E-commerce

Aus daten-speicherung.de
Zur Navigation springen Zur Suche springen

Draft submission for public consultation on the future of electronic commerce in the internal market and the implementation of the Directive on electronic commerce (2000/31/EC) that is open until 5 November 2010:

Accessibility of company data protection officer

It is becoming more and more important that queries regarding personal data can be addressed directly to a company data protection officer where such a person has been appointed. Service providers should therefore provide the details of a company data protection officer, including his electronic mail address, which allow him to be contacted rapidly and communicated with in a direct and effective manner (article 5 ECD to be amended).

Intermediary liability: Protect freedom of speech by stopping pro-active monitoring and "prevention"

Case study: Inacceptable situation in Germany

The German Federal Supreme Court (Bundesgerichtshof)[1] interprets the liability exemptions in the e-commerce directive not to be cover applications for injunctive relief, i.e. to claims for providers to prevent illegal user activity.[2].

In consequence, the most far-reaching doctrine of contributory "liability" is being applied by German courts, called "accessory liability" (Störerhaftung). According to this doctrine it is not only the wrongdoer himself (direct infringer) and participants (effective promoters or helpers) that can be subject to a claim for refraining from rights infringements, but also mere accessories, including providers of information society services. Responsibility for unlawful user action is extended to all persons who - without necessarily being wrongdoers or participants - deliberately and generally causally contribute to the infringement of a third party's right, provided they have the legal and effective means to detect and prevent the infringement.[3]

According to this doctrine, once an intermediary obtains knowledge of an infringement, it is not only obliged to remove the unlawful content but also to take all technically feasible and reasonable precautions to prevent future infringements. In other words, subject to the requirement of reasonableness, service providers are obliged to examine all user content as soon as they obtain knowledge of any unlawful content. The monitoring obligation is not limited to the detection of the unlawful content that was originally notified or to the original publisher of this content.[4]

In practice, this jurisprudence effectively forces providers to pro-actively monitor user-generated content, this being the only way they can avoid indictment.[5] Culpable breach of the monitoring duty is punished by a disciplinary fine or even a prison sentence.

A German court that finds a provider not to have taken sufficient steps will grant injunctive relief by prohibiting the provider to allow users to re-publish the illegal content, without setting out what measures the court considers necessary to prevent user infringements.[6] The extent of the provider's pro-active obligations is determined only when the rights holder initiates a separate procedure for alleged violation of the injunction and applies for a disciplinary fine or a prison sentence to be imposed. So providers are told which steps to take only in the judgement that imposes a fine on them. This jurisprudence causes unacceptable uncertainty for providers.[7] It is impossible for intermediaries to anticipate which measures the courts would consider "reasonable".[8]

This doctrine also has unacceptable repercussions on the freedom of speech on-line, because providers threatened by fines will - often using automatic and broad filters - block and prevent any content that they consider a risk. This leads to the suppression of controversial but politically very valuable content, as intermediaries to not wish the battle on the legality of the content to be fought at their expense.

Furthermore, German courts tend to go very far in what measures they impose on providers to prevent future infringements by their users: It was held that there was a duty to deploy technology to automatically scan all user-generated content for content that potentially violates an injunction.[9] Users that have generated illegal content in the past must be subjected to a manual examination of any future content they generate.[10] To this end, all users must be identified and the anonymous use of services must be disabled.[11] All user actions must be logged.[12] Content regarding issues that "provoke illegal reaction" must be examined manually.[13] All of these duties have been imposed not for the prevention of serious crime, but for the mere prevention of infringements of private titles including commercial rights.

The concept of private policing is failed

These "preventive duties" imposed by the courts violate article 15 of the e-commerce directive, according to which Member States - including their courts - must not impose a general obligation on providers to monitor information which they transmit or store. Injunctions lead to a de facto obligation to monitor user-generated content[14], and thus amount to a general monitoring obligation.

At any rate, the entire concept of "specific" obligations to "prevent" illegal user action is failed. In a democracy, everybody can trust in the integrity of their fellow citizens, even if it is known that this trust is sometimes abused. Fundamental freedoms constitute the foundation of justice and peace in the world.[15] Freedom is the purpose of all law. Its benefits for every person as well as for our society as a whole by far outweigh the harm done by its abuse. A prevention society aimed at eliminating, as far as possible, all potential risks of human behaviour and life is not compatible with European values and freedoms.

According to general principles of civil law, only those need to prevent harm that have created a hazard source. The exchange of information is at the roots of human nature. It is a fundamental right (article 11 of the charter of fundamental rights) and can therefore not be considered a "hazard source". The provision of telecommunications services does not create a greater risk of rights infringements than the provision of any product or service. Typical, socially adequate and therefore legal risks do not put their originator in a position of being responsible for intentional violations committed by other people.

Even prevention (or policing) technology that can reasonably be implemented or is industry standard must not be imposed on all service providers because of its devastating effects on freedom of speech. Filtering technology, for example, will by its nature suppress legal content that is merely similar to illegal content. This leads to the suppression of controversial but politically very valuable content, for example critical comments on companies and products or "fair use" of intellectual property. Policing is not the job of private companies.

Recommendation

The concept of "specific" obligations of intermediaries to "prevent" illegal user action that is rooted in article 14 (3) and recitals 47 and 48 of the e-commerce directive should thus be given up in its entirety. Intermediaries should only be required to remove illegal content upon notification. The deterring effect of criminal sanctions is sufficient to "prevent" illegal user action.

Recommendation:

  • Service providers should not be required to "prevent" infringements. To this end, article 14 (3) ECD should read: "This Article shall not affect the possibility for a court or administrative authority, in accordance with Member States' legal systems, of requiring the service provider to terminate or prevent an infringement, nor does it affect the possibility for Member States of establishing procedures governing the removal or disabling of access to information." Recitals 47 and 48 should be deleted.

Intermediary liability: Judicial review is needed to protect freedom of speech

The current liability limitations do not sufficiently protect free speech for another reason: A service provider that is notified of allegedly illegal content is not exempted from liability under the current directive even if is has reason to believe that the content may well be legal. In practise this situation leads to the removal of practically any notified content without proper assessment of its legality, resulting in major damage to free speech on-line. The provider is being put in the position of a judge which it cannot fill. If it removes content which is later considered legal by the courts, it can be faced with high damage claims by its customer/user. If the provider refuses to remove content which is later considered illegal by the courts, it can be faced with high damage claims by the rights holder.

A service provider that is notified of allegedly illegal content should therefore not be required to remove the content before its legality has been assessed by a judge in a preliminary procedure. Member States can design this procedure to be fast and effective. However its cost must not be borne by the provider as this would again have a chilling effect on free speech. Needy claimants can use legal aid. Claimants can recover legal expenses by suing the user that generated the content.

Recommendation:

  • Service providers should not be required to remove or disable access to information before it has been found illegal by a court of law. To this end, article 14 (1) (b) ECD should read: "the provider, upon obtaining such knowledge or awareness, acts expeditiously to remove or to disable access to the information after it has been found illegal by a court of law."

Internet anonymity and privacy

On the Internet our every action can be tracked easily and comprehensively. Also information that was collected on the Internet is routinely lost or stolen. This situation has caused many potential users to refrain from using information society services, or even from using the Internet.

For example a German poll found that 50% of the polled have "often" refrained from on-line orders because they did not want to provide personal information that was required.[16] Even 40% of the 19 to 29 year olds confirm this. 18% even say they never order anything on the Internet because they do not want to provide personal information. 12% of the 19 to 29 year olds confirmed this.

The collection of unnecessary personal information on the Internet thus constitutes an obstacle to economic development and employment in Europe. Looking for short-term profits, companies are ignoring that their user screening practises are eroding consumer trust and the basis of the long-term success of e-commerce.

For those reasons it is necessary to establish strict sector-specific privacy rules for information society services:

  1. The provider of an information society service shall offer the user anonymous use and payment of information society services to the extent technically feasible and reasonable. The provider shall offer the user anonymous use and payment of teleservices in particular where services of this kind are already being offered anonymously by competitors. The user shall be informed about these options.
  2. The provider of an information society service may collect, process and use personal data concerning the use of of an information society services only to the extent necessary to enable the user to use the information society service (transactional data) or to charge the user for the use of the information society service.
  3. The provider shall ensure that personal data generated in connection with the process of using information society services are erased or rendered anonymous as soon as possible, at the latest upon the end of each utilization, unless further storage is required for billing purposes.
  4. The provider may not make the provision of an information society service dependent on the supply of personal data that are not required for the provision of the service.
  5. The provider must not make the provision of an information society service dependent on the consent of the data subject to the processing or use of their data for other purposes than the provision of the service.
  6. Directive 93/13 of 5 April 1993 on unfair terms in consumer contracts should be extended to cover consenting clauses concerning personal data which have not been individually negotiated.
  7. The provider must reveal the periods of time for which specific types of personal data are typically retained.

Recommendation:

  • The e-commerce directive should be amended to include strict sector-specific privacy rules for information society services.

Non-discrimination

Users that exercise data protection rights and attempt to end illegal data collection and processing practises are often silenced by terminating their account. This way providers can avoid having to comply with privacy laws.

This situation has devastating effects not only for the user whose account is closed, but for all other users of that service whose rights cannot be enforced and also for competitors that respect privacy laws and are unfaily disadvantaged in comparison to the non-complying provider.

Recommendation:

  • The provider may not discriminate against a user that legally exercises data protection rights. The e-commerce directive should be amended to this effect.

Confidentiality of our Internet use

The confidentiality of our Internet use must be protected not only by the access provider, but also by providers of information society services. The e-commerce directive should be amended to include a similar provision to article 5 of directive 2002/58/EC.

Recommendation:

  • The e-commerce directive should be amended as follows: "Member States shall ensure the confidentiality of the usage of information society services and the related traffic and suscriber data, through national legislation. In particular, they shall prohibit listening, tapping, storage or other kinds of interception or surveillance of usage and the related traffic and subscriber data by persons other than users, without the consent of the users concerned, except when legally authorised to do so. This paragraph shall not prevent technical storage which is necessary for the provision of a service without prejudice to the principle of confidentiality."


(More information in German)

  1. BGH, NJW 2004, 3102 (3103)
  2. http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/germany_12nov2007_en.pdf
  3. DG Market, Study on liability of Internet providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 51.
  4. DG Market, Study on the liability of Internet intermediaries, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/germany_12nov2007_en.pdf.
  5. DG Market, Study on liability of Internet providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 51.
  6. DG Market, Study on liability of Internet providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 66.
  7. DG Market, Study on liability of Internet service providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 67.
  8. DG Market, Study on liability of Internet providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 51.
  9. BGH, NJW 2007, 2636 (2639 f.)
  10. BGH, NJW 2008, 758 (762) - eBay
  11. OLG Hamburg, Urteil vom 02.07.2008 – 5 U 73/07 – Rapidshare.
  12. OLG Hamburg, Urteil vom 02.07.2008 – 5 U 73/07 – Rapidshare.
  13. OLG Hamburg, Urteil vom 22.08.2006 – 7 U 50/06 – Heise-Forum.
  14. DG Market, Study on liability of Internet providers, http://ec.europa.eu/internal_market/e-commerce/docs/study/liability/final_report_en.pdf, p. 51.
  15. European Convention on Human Rights, preamble.
  16. Allensbach study in August of 2010, http://www.webcitation.org/5t9uDMnHb