Government's AI Copyright Consultation is Selling out to the Techbros
We have recently seen the publication of the Government's Copyright and AI Consultation paper. This my take on it.
I co-chair the All Party Parliamentary Group for AI and chaired the AI select Committee committee and wrote a book earlier this year on AI regulation. Before that I had a career as an lawyer defending copyright and creativity and in the House of Lords, I’ve have been my Party’s creative industry spokesperson. The question of IP and AI absolutely for me is the key issue which has arisen in relation to Generative AI models. It is one thing to use tech, another to be at the mercy of it.
It is a major issue not just in the UK, but around the world. Getty and the New York Times are suing in the United States, so too many writers, artists and musicians and it was at the root of the Hollywood Actor and Writers strike last year .
Here in the UK, as the Government’s intentions have become clearer the temperature has risen. We have seen the creation of a new campaign -Creative Rights in AI Coalition (CRAIC) across the creative and news industries and Ed Newton-Rex raising over 30,000 signatories from creators and creative organisations.
But with the new government consultation which came out a few days ago we are now faced with a proposal regarding text and data mining exception which we thought was settled under the last Government. It starts from the false premise of legal uncertainty.
As the News Media Association say:
The government’s consultation is based on the mistaken idea—promoted by tech lobbyists and echoed in the consultation—that there is a lack of clarity in existing copyright law. This is completely untrue: the use of copyrighted content by Gen AI firms without a license is theft on a mass scale, and there is no objective case for a new text and data mining exception.
There is no lack of clarity over how AI developers can legally access training data. UK law is absolutely clear that commercial organisations – including Gen AI developers – must license the data they use to train their Large Language Models (“LLMs”).
Merely because AI platforms such as Stability AI are resisting claims doesn’t mean the law in the UK is uncertain. There is no need for developers to find ‘it difficult to navigate copyright law in the UK’.
AI developers have already in a number of cases reached agreement with between news publishers. OpenAI has signed deals with publishers like News Corp, Axel Springer, The Atlantic, and Reuters, offering annual payments between $1 million and $5 million, with News Corp’s deal reportedly worth $250 million over five years.
There can be no excuse of market failure. There are well established licensing solutions administered by a variety of well-established mechanisms and collecting societies. There should be no uncertainty around the existing law. We have some of the most effective collective rights organisations in the world. Licensing is their bread and butter.
The Consultation paper says that “The government believes that the best way to achieve these objectives is through a package of interventions that can balance the needs of the two sectors” Ministers Lord Vallance, and Feryal Clark MP seem to think we need a balance between the creative industries and the tech industries. But what kind of balance is this?
The government is proposing to change the UK’s copyright framework by creating a text and data mining exception where rights holders have not expressly reserved their rights—in other words, an ‘opt-out’ system, where content is free to use unless a rights holder proactively withholds consent. To complement this, the government is proposing: (a) transparency provisions; and (b) provisions to ensure that rights reservation mechanisms are effective.
The government has stated that it will only move ahead with its preferred ‘rights reservation’ option if the transparency and rights reservation provisions are ‘effective, accessible, and widely adopted’. However, it will be up to Ministers to decide what provisions meet this standard, and it is clear that the government wishes to move ahead with this option regardless of workability, without knowing if their own standards for implementation can be met.
Although it is absolutely clear that we know that use of copyright works to train AI models is contrary to UK copyright law, the laws around transparency of these activities haven’t caught up. As well as using pirated e-books in their training data, AI developers scrape the internet for valuable professional journalism and other media in breach of both the terms of service of websites and copyright law, for use in training commercial AI models.
At present, developers can do this without declaring their identity, or they may use IP scraped to appear in a search index for the completely different commercial purpose of training AI models.
How can rights owners opt-out of something they don’t know about? AI developers will often scrape websites, or access other pirated material before they launch an LLM in public. This means there is no way for IP owners to opt-out of their material being taken before its inclusion in these models. And once used to train these models, the commercial value has already been extracted from IP scraped without permission with no way to delete data from those models.
The next wave of AI models responds to user queries by browsing the web to extract valuable news and information from professional news websites. This is known as Retrieval Augmented Generation-RAG. Without payment for extracting this commercial value, AI agents built by companies such as Perplexity, Google and Meta, will effectively free ride on the professional hard work of journalists, authors and creators. At present such crawlers are hard to block.
This is incredibly concerning, given that no effective ‘rights reservation’ system for the use of content by Gen AI models has been proposed or implemented anywhere in the world, making the government proposals entirely speculative.
As the NMA also say What the government is proposing is an incredibly unfair trade-off—giving the creative industries a vague commitment to transparency, whilst giving the rights of hundreds of thousands of creators to Gen AI firms. While creators are desperate for a solution after years of copyright theft by Gen AI firms, making a crime legal cannot be the solution to mass theft.
We need transparency and clear statement about copyright. We absolutely should not expect artists to have to opt out. AI developers must: be transparent about the identity of their crawlers; be transparent about the purposes of their crawlers; and have separate crawlers for distinct purposes.
Unless news publishers and the broader creative industries can retain control over their data – making UK copyright law enforceable – AI firms will be free to scrape the web without remunerating creators. This will not only reduce investment in trusted journalism, but it will ultimately harm innovation in the AI sector. If less and less human-authored IP is produced, tech developers will lack the high-quality data that is the essential fuel in generative AI.
Amending UK law to address the challenges posed by AI development, particularly in relation to copyright and transparency, is essential to protect the rights of creators, foster responsible innovation, and ensure a sustainable future for the creative industries.
This should apply regardless of which country the scraping of copyright material takes place if developers market their product in the UK, regardless of where the training takes place.
It will also ensure that AI start-ups based in the UK are not put at a competitive disadvantage due to the ability of international firms to conduct training in a different jurisdiction
It is clear that AI developers have used their lobbying clout to persuade the government that a new exemption from copyright in their favour is required. As a result, the government seem to have sold out to the tech bros.
In response the creative industries and supporters such as myself will be vigorously opposing government plans for a new text and data mining exemption and ensuring we get answers to our questions:
What led the government to do a u-turn on the previous government’s decision to drop the text and data mining exemption it proposed?
What estimate of the damage to the creative industries it has made of implementing its clearly favoured option of a TDM plus opt out?
Is damaging the most successful UK economic sector for the benefit of US AI developers what it means by balance?
Why it has not included the possibility of an opt in to a TDM in its consultation paper options?
What is the difference between rights reservation and opting out? Isn’t this pure semantics?
What examples of successful workable opt outs or rights reservation from TDM’s can it draw on particularly for small rights holders? What research has it done? the paper essentially admits that effective technology is not there yet. Isn’t it clear that the EU opt out system under the Copyright Directive has not delivered clarity?
What regulatory mechanism if any does the government envisage if its proposal for a TDM with rights reservation/opt out is adopted? How are creators going to be sure any new system would work in the first place?
We Need Better Protection for Citizens in the Face of Automated Decision Making
The second Reading of my Private Members Bill tool place recently. It is designed to give greater rights to all of us who are subject to AI and Automated decision making in government which is becoming increasingly prevalent with the enthusiasm of the new Labour government to "digitally transform" our public services.
I thank Big Brother Watch, the Public Law Project and the Ada Lovelace Institute, which, each in their own way, have provided the evidence and underpinned my resolve to ensure that we regulate the adoption of algorithmic and AI tools in the public sector, which are increasingly being used across it to make and support many of the highest-impact decisions affecting individuals, families and communities across healthcare, welfare, education, policing, immigration and many other sensitive areas of an individual’s life. I also thank the Public Bill Office, the Library and other members of staff for all their assistance in bringing this Bill forward and communicating its intent and contents, and I thank all noble Lords who have taken the trouble to come to take part in this debate this afternoon.
The speed and volume of decision-making that new technologies will deliver is unprecedented. They have the potential to offer significant benefits, including improved efficiency and cost effectiveness in government operations, enhanced service delivery and resource allocation, better prediction and support for vulnerable people and increased transparency in public engagement. However, the rapid adoption of AI in the public sector also presents significant risks and challenges, with the potential for unfairness, discrimination and misuse through algorithmic bias and the need for human oversight, a lack of transparency and accountability in automated decision-making processes and privacy and data protection concerns.
Incidents such as the 2020 A-level and GCSE grading fiasco, where an algorithmic approach saw students, particularly those from lower-income areas, unfairly miss out on university places when an algorithm was used to estimate grades from exams that were cancelled because of Covid-19, have starkly illustrated the dangers of unchecked algorithmic systems in public administration disproportionately affecting those from lower-income backgrounds. That led to widespread public outcry and a loss of trust in government use of technology.
Big Brother Watch’s investigations have revealed that councils across the UK are conducting mass profiling and citizen scoring of welfare and social care recipients. Its report, entitled Poverty Panopticon [The Hidden Algorithms Shaping Britain’s Welfare State], uncovered alarming statistics. Some 540,000 benefits applicants are secretly assigned fraud risk scores by councils’ algorithms before accessing housing benefit or council tax support. Personal data from 1.6 million people living in social housing is processed by commercial algorithms to predict rent non-payers. Over 250,000 people’s data is processed by secretive automated tools to predict the likelihood of abuse, homelessness or unemployment.
Big Brother Watch criticises the nature of these algorithms, stating that most are secretive, unevidenced, incredibly invasive and likely discriminatory. It argues that these tools are being used without residents’ knowledge, effectively creating tools of automated suspicion. The organisation rightly expressed deep concern that these risk-scoring algorithms could be disadvantaging and discriminating against Britain’s poor. It warns of potential violations of privacy and equality rights, drawing parallels to controversial systems like the Metropolitan Police’s gangs matrix database, which was found to be operating unlawfully. From a series of freedom of information requests last June, Big Brother Watch found that a flawed DWP algorithm wrongly flagged 200,000 housing benefit claimants for possible fraud and error, which meant that thousands of UK households every month had their housing benefit claims unnecessarily investigated.
In August 2020, the Home Office agreed to stop using an algorithm to help sort visa applications after it was discovered that the algorithm contained entrenched racism and bias, and following a challenge from the Joint Council for the Welfare of Immigrants and the digital rights group Foxglove. The algorithm essentially created a three-tier system for immigration, with a speedy boarding lane for white people from the countries most favoured by the system. Privacy International has raised concerns about the Home Office's use of a current tool called Identify and Prioritise Immigration Cases—IPIC—which uses personal data, including biometric and criminal records to prioritise deportation cases, arguing that it lacks transparency and may encourage officials to accept recommended decisions without proper scrutiny.
Automated decision-making has been proven to lead to harms in privacy and equality contexts, such as in the Harm Assessment Risk Tool, which was used by Durham Police until 2021, and which predicted reoffending risks partly based on an individual’s postcode in order to inform charging decisions. All these cases illustrate how ADM can perpetuate discrimination. The Horizon saga illustrates how difficult it is to secure proper redress once the computer says no.
There is no doubt that our new Government are enthusiastic about the adoption of AI in the public sector. Both the DSIT Secretary of State and Feryal Clark, the AI Minister, are on the record about the adoption of AI in public services. They have ambitious plans to use AI and other technologies to transform public service delivery. Peter Kyle has said:
“We’re putting AI at the heart of the government’s agenda to boost growth and improve our public services”,
and
“bringing together digital, data and technology experts from across Government under one roof, my Department will drive forward the transformation of the state”.—[Official Report, Commons, 2/9/24; col. 89.]
Feryal Clarke has emphasised the Administration’s desire to “completely transform digital Government” with DSIT. As the Government continue to adopt AI technologies, it is crucial to balance the potential benefits with the need for responsible and ethical implementation to ensure fairness, transparency and public trust.
The Ada Lovelace Institute warns of the unintended consequences of AI in the public sector, including the risk of entrenching existing practices, instead of fostering innovation and systemic solutions. As it says, the safeguards around automated decision-making, which exist only in data protection law, are therefore more critical than ever in ensuring people understand when a significant decision about them is being automated, why that decision is made, and have routes to challenge it, or ask for it to be decided by a human.
Our citizens need greater, not less, protection, but rather than accepting the need for these, we see the Government following in the footsteps of their predecessor by watering down such rights as there are under GDPR Article 22 not to be subject to automated decision-making. We will, of course, be discussing these aspects of the Data (Use and Access) Bill in Committee next week.
ADM safeguards are critical to public trust in AI, but progress has been glacial. Take the Algorithmic Transparency Recording Standard, which was created in 2022 and is intended to offer a consistent framework for public bodies to publish details of the algorithms used in making these decisions. Six records were published at launch, and only three more seem to have been published since then. The previous Government announced earlier this year that the implementation of the Algorithmic Transparency Recording Standard will be mandatory for departments. Minister Clark in the new Government has said,
“multiple records are expected to be published soon”,
but when will this be consistent across government departments? What teeth do the Central Digital and Data Office and the Responsible Technology Adoption Unit, now both within DSIT, have to ensure the adoption of the standard, especially in view of the planned watering down of the Article 22 GDPR safeguards? Where is the promised repository for ATRS records? What about the other public services in local government too?
The Public Law Project, which maintains a register called Tracking Automated Government, believes that in October last year there were more than 55 examples of public ADM systems use. Where is the transparency on those? The fact is that the Government’s Algorithmic Transparency Recording Standard, while a step in the right direction, remains voluntary and lacks comprehensive adoption or indeed a compliance mechanism or opportunity for redress. The current regulatory landscape is clearly inadequate to address these challenges. Despite the existing guidance and framework, there is no legally enforceable obligation on public authorities to be transparent about their use of ADM and algorithmic systems, or to rigorously assess their impact.
To address these challenges, several measures are needed. We need to see the creation of and adherence to ethical guidelines and accountability mechanisms for AI implementation; a clear regulatory framework and standards for use in the public sector; increased transparency and explainability of the adoption and use of AI systems; investment in AI education; and workforce development for public sector employees. We also need to see the right of redress, with a strengthened right for the individuals to challenge automated decisions.
My Bill aims to establish a clear mandatory framework for the responsible use of algorithmic and automated decision-making systems in the public sector. It will help to prevent the embedding of bias and discrimination in administrative decision-making, protect individual rights and foster public trust in government use of new technologies.
I will not adumbrate all the elements of the Bill. In an era when AI and algorithmic systems are becoming increasingly central to government ambitions for greater productivity and public service delivery, this Bill, I hope noble Lords agree, is crucial to ensuring that the benefits of these technologies are realised while safeguarding democratic values and individual rights. By ensuring that ADM systems are used responsibly and ethically, the Bill facilitates their role in improving public service delivery, making government operations more efficient and responsive.
The Bill is not merely a response to past failures but a proactive measure to guide the future use of technology within government and empower our citizens in the face of these powerful new technologies. I hope that the House and the Government will agree that this is the way forward.
Lords Debate Regulators : Who Watches the Watchdogs?
Recently the Lords held a debate on the report of the Industry and Regulators Select Committee (on which I sit) entitled "Who Watches the Watchdogs" about the scrutiny given to the performance independence and competence of our regulators
This is what I said. It was an opportunity as ever to emphasize that Regulation is not the enemy of innovation, or indeed growth, but can in fact, by providing certainty of standards, be the platform for it.
The Grenfell report and today’s Statement have been an extremely sobering reminder of the importance of effective regulation and the effective oversight of regulators. The principal job of regulation is to ensure societal safety and benefit—in essence, mitigating risk. In that context, the performance of the UK regulators, as well as the nature of regulation, is crucial.
In the early part of this year, the spotlight was on regulation and the effectiveness of our regulators. Our report was followed by a major contribution to the debate from the Institute for Government. We then had the Government’s own White Paper, Smarter Regulation, which seemed designed principally to take the growth duty established in 2015 even further with a more permissive approach to risk and a “service mindset”, and risked creating less clarity with yet another set of regulatory principles going beyond those in the Better Regulation Framework and the Regulators’ Code.
Our report was, however, described as excellent by the Minister for Investment and Regulatory Reform in the Department for Business and Trade under the previous Government, the noble Lord, Lord Johnson of Lainston, whom I am pleased to see taking part in the debate today. I hope that the new Government will agree with that assessment and take our recommendations further forward.
Both we and the Institute for Government identified a worrying lack of scrutiny of our regulators—indeed, a worrying lack of even identifying who our regulators are. The NAO puts the number of regulators at around 90 and the Institute for Government at 116, but some believe that there are as many as 200 that we need to take account of. So it is welcome that the previous Government’s response said that a register of regulators, detailing all UK regulators, their roles, duties and sponsor departments, was in the offing. Is this ready to be launched?
The crux of our report was to address performance, strategic independence and oversight of UK regulators. In exploring existing oversight, accountability measures and the effectiveness of parliamentary oversight, it was clear that we needed to improve self-reporting by regulators. However, a growth duty performance framework, as proposed in the White Paper, does not fit the bill.
Regulators should also be subject to regular performance evaluations, as we recommended; these reviews should be made public to ensure transparency and accountability. To ensure that these are effective, we recommended, as the noble Lord, Lord Hollick mentioned, establishing a new office for regulatory performance—an independent statutory body analogous to the National Audit Office—to undertake regular performance reviews of regulators and to report to Parliament. It was good to see that, similar to our proposal, the Institute for Government called for a regulatory oversight support unit in its subsequent report, Parliament and Regulators.
As regards independence, we had concerns about the potential politicisation of regulatory appointments. Appointment processes for regulators should be transparent and merit-based, with greater parliamentary scrutiny to avoid politicisation. Although strategic guidance from the Government is necessary, it should not compromise the operational independence of regulators.
What is the new Government’s approach to this? Labour’s general election manifesto emphasised fostering innovation and improving regulation to support economic growth, with a key proposal to establish a regulatory innovation office in order to streamline regulatory processes for new technologies and set targets for tech regulators. I hope that that does not take us down the same trajectory as the previous Government. Regulation is not the enemy of innovation, or indeed growth, but can in fact, by providing certainty of standards, be the platform for it.
At the time of our report, the IfG rightly said:
“It would be a mistake for the committee to consider its work complete … new members can build on its agenda in their future work, including by fleshing out its proposals for how ‘Ofreg’ would work in practice”.
We should take that to heart. There is still a great deal of work to do to make sure that our regulators are clearly independent of government, are able to work effectively, and are properly resourced and scrutinised. I hope that the new Government will engage closely with the committee in their work.
Lord C-J Commentary on the new Government's Science and Technology Programme
Sadly we only had 5 minutes speaking time in the recent Kings Speech debate . Here is an an extended version of my speech which goes into greater depth as to what I believe the Government should be doing in this area if it is to fulfill its growth through innovation agenda and expresses some caveats about how they plan to do this.
When we debated the New government’s proposals in the Kings speech recently the House of Lords gave a particularly warm welcome to Lord Vallance of Balham-formerly Sir Patrick Vallance- as the new Minister of State in the Department. While the Government’s Chief Scientific Adviser we know from the book “the Long Shot” how he played an critical role in the establishment of the UK Vaccine Taskforce, which was set up in April 2020 in response to the COVID-19 pandemic. He was pivotal in the recruitment of Dame Kate Bingham to chair the Vaccine Taskforce and in organizing the overall strategy for the UK development and distribution of COVID-19 vaccines. For that we should be eternally grateful.
I welcome the Government’s growth through innovation agenda and mission to enhance public services through the deployment of new technology and also the concentration of digital functions in DSIT and that it will become the centre for “digital expertise and delivery in government,improving how the government and public services interact with citizens.” in the words of the new Secretary of State, Peter Kyle.
The Government is expanding the department’s scope and size by bringing in experts in data, digital, and AI from the Government Digital Service, the Incubator for AI (i.AI), and the Central Digital Data Office to unite efforts to implement digital transformation of public services under one roof. There is great potential in justice, education, healthcare to name but three areas.
This is crucial particularly in the adoption of innovative technologies and tools in our healthcare for which Liberal Democrats believe there should be ring-fenced budgets. We need to be ensuring interoperability of IT systems too.
They government have committed too to modernising public sector procurement frameworks to enable start-ups and SMEs to drive public sector innovation and better public services. Will , however, clear, transparent framework of standards incorporating ethical principles be established? Public sector adoption is very desirable but requires trust on the part of the public/ and the citizen For instance we need to ensure that citizens can assert their rights when faced with automated decision making or live facial recognition
It has felt, under the previous regime, that universities have been under continual threat from government rather than valued as the engines of knowledge and growth and we need to be far more internationally outward looking, in particular fixing our relationship with the EU- using science and technology to address societal challenges for a more resilient and prosperous future in the words of the Royal Society.
I welcome the new Industrial Strategy Council. Does this mean we can plan for 10 years of stability and opportunity creation in science and tech sector? Successive policy changes to the R&D tax regime over the past several years have created uncertainty and additional red-tape for SMEs, putting at risk the UK’s reputation as a location for innovative businesses.We need to give businesses certainty and incentivise them to invest in new technologies to grow the economy, create good jobs and tackle the climate crisis.
Opening up what can be a blocked pipeline all the way from R & D to commercialisation, from university spinout through start up to scale up and IPO, and crowding in and derisking private investment through the National Wealth Fund, the British Business Bank and post Mansion House pension reforms, are crucial with all the local, regional, national and UK wide aspects, recognizing the importance of innovation clusters and centres of excellence. We need to tackle regional disparities and develop the innovation clusters with greater devolution to combined authorities
Digital Skills and Digital literacy are also crucial but to deploy digital tools successfully we also need a pipeline of creative collaborative and critical thinking skills. A massive skills and upskilling agenda is needed in the face of technology advances. The focus in training should be on lifelong skills grants, reforming the apprenticeship levy, and boosting vocational training and apprenticeships and many of the governments proposals in this respect are welcome.
In this context, as the the chair of a university governing council I very much welcome the Government’s new tone on the value of universities, of long term settlements, and of resetting relations with Europe and international research collaboration.
The role of university research and spinouts is crucial . The Research Excellence Framework has the perverse incentive of discouraging cooperation. We should be encouraging strategic partnerships in research especially internationally. We need to be full throated members of Horizon -the uncertainty has been extremely damaging to collaboration. I hope the government will now commit to joining the European Innovation Council as well
Last year Labour set out its plan for the life sciences.It committed to the investment of £10bn into R&D. Further, the plan said that Labour would see the creation of 100,000 jobs in the life sciences sector by 2030. The document contains a range of further welcome pledges including strengthening the Office for Life Sciences and the Life Sciences Council, and to bring laboratory clusters within the scope of the ‘Nationally significant infrastructure regime’ in England.
We need to ensure Government spending on R&D keeps pace with other nations, and establish a long-term strategy for science, research and innovation that commands cross party support.Research, development and innovation are crucial to driving productivity growth, yet our current levels of R&D investment and productivity lag the G7. I hope this means that we will soon see whether spending plans for government R & D expenditure by 2030 and 2035 match their words.
And disproportionately high overseas researcher visa costs MUST be lowered as Lord Vallance recommended in his Digital Technology Review. UK visa costs are up to 17 times higher than other leading science nations.The Royal Society have called this a “punitive tax on talent”.
But support for innovation should not be unconditional or at any cost. I hope this government will not fall into the trap of viewing regulation as necessarily the enemy of innovation. We need guardrails to ensure that, for example, AI adoption leads to public benefit.
I hope therefore that the reference to AI regulation in the King's Speech, but failure to announce a bill, is only a timing issue. What IS the Government’s intention especially given an AI bill was heavily trailed in the media?
With AI technologies continuing to develop at an exponential rate, clarity on regulation is needed by developers and adopters.There is the question too as to what extent the new government will depart from the current sectoral approach to regulating AI and adopt a cross-sectoral approach. What does the King's Speech reference to regulating "the most powerful artificial intelligence models" actually refer to? Will the government be launching yet another consultation on AI regulation?
There is no doubt we need to seize the opportunities of AI, whilst making sure we mitigate the risks of AI, ensuring ethical standards for AI development and use are adopted.
Liberal Democrats believe we need to create a clear, workable and well-resourced cross-sectoral regulatory framework for artificial intelligence that:
- Promotes innovation while creating certainty for AI users, developers and investors.
- Establishes transparency and accountability for AI systems in the public sector.
- Ensures the use of personal data and AI is unbiased, transparent and accurate, and respects the privacy of innocent people
The government in particular should lead the way in ensuring that there is a high level of transparency and opportunity for redress when Algorithmic and automated systems are used in government. I commend my new private members bill (the Public Authority Algorithmic and Automated Decision-Making Systems Bill) to it!
The government should also negotiate the UK’s participation in the Trade and Technology Council with the US and the EU, so we can play a leading role in global AI regulation, and we should work with international partners in agreeing common global standards for AI risk and impact assessment, testing, training monitoring and audit.
As regards AI regulation in the Kings Speech itself we are promised a Product Safety and Metrology bill which could require alignment of AI driven products with the EU AI Act which seems to be putting the cart well in front of the AI regulatory horse.
We do need however to ensure that high risk systems are mandated to adopt international ethical and safety standards.At the same time in In this age of IOT we should require all suppliers to provide a short, clear version of their terms and conditions, setting out the key facts as they relate to individuals’ data and privacy.
As regards the creative industries there are clearly great opportunities in relation to the use of AI but there are also challenges and big questions over authorship and intellectual property and many artists feel threatened-the root cause of the recent Hollywood writers and actors strike. What is the government’s approach?
We need to establish very clearly that Generative AI systems need a licence to ingest copyright material for training purposes-just as Mumsnet and the New York Times are asserting- and that there is an obligation of transparency in the use of data sets and original content.
Lord Vallance is on record as wanting certainty in the relationship between IP rights and generative AI for innovator and investor confidence. And this should be the case for for creatives too. Copyright content needs to be properly remunerated by the tech platforms. The bill needs to make clear that platforms profit from content and need to pay properly and fairly, on benchmarked terms and with reference to value for end users when content is use for training Large Language Models.
And when will the government set up the promised new Regulatory Innovation Office? This was promised as an organisation to help “regulators to update regulation, speed up approval timelines and co-ordinate issues that span existing boundaries”. and as a “pro-innovation body” designed to “set targets for tech regulators, end uncertainty for businesses, turbocharge output, and boost economic growth”. We need in particular to know whether it will replace the Digital Regulators Cooperation Forum.
We must also ensure we have the right climate for FDI. The Harrington Report called for a new Business investment Strategy for the Office for Investment. Despite the previous government’s Life Sciences Vision we have seen pharma company Eli Lilley pulling investment on laboratory space in London because the UK “does not invite inward investment at this time”. Astra Zeneca decided to build its next plant in Ireland because of the U.K.’s “discouraging” tax rate.
We also need to modernise employment rights to make them fit for the age of the gig economy,including by establishing a new ‘dependent contractor’ employment status in between employment and self-employment, with entitlements to basic rights such as minimum earnings levels, sick pay and holiday entitlement.
There is a great need for need for greater diversity and inclusion in the AI workforce and science and technology more broadly. Only one in four senior tech employees in the UK are women, and only 14% from ethnic minorities.
I hope the Government too is fully committed despite its growth agenda to a full hearted support for the Competition and Markets Authority in the use of its powers under the new Digital Markets Act. I welcome the CMA’s market investigation into Cloud Services and its reassurance that it is looking broadly at the anti-competitive practices of the service providers such as vendor lock-in tactics and non-competitive procurement.
Then again how will the government kickstart better progress on Project Gigabit? Given the competitive model for rollout of broadband services that has been chosen, investors in alternative providers to the incumbents need reassurance that their investment is going onto a level playing field and not one tilted in favour of the incumbents.
Also in terms of vital cross departmental working, joining up government on Science and Technology policy we need to know what the role will be of the National Science and Technology Council and what are its key priorities.
There no mention in Labour’s manifesto on the potential impact of AI on the workplace.The TUC and Institute for the Future of Work are among those who have called for new legislation to create further legal protections for workers and employers in relation to the use of AI. The government should introduce safeguards against the invasion of privacy through surveillance technology and discriminatory algorithmic decision-making in the workplace along the lines of the TUC draft bill and algorithmic impact assessment along the lines of IFOW’s proposals.
The Government’s will also need to decide how to follow up on the recommendations of recent key Reports such as
- Professor Dame Angela McLean’s Review of Life Sciences
- The Vallance Review of Pro-innovation Regulation of Digital Technologies
- The Independent Review of Research Bureaucracy by Professor Adam Tickell
- The Independent Review of the UKRI by Sir David Grant
- The Independent Review of the UK’s Research, Development and Innovation Landscape by Sir Paul Nurse
- The O’Shaughnessy Report on Clinical Trials
- The Independent Review of the Future of Compute by Professor Zoubin Ghahramani FRS and
- The Independent Review of University Spin-out Companies by Professor Irene Tracey and Dr. Andrew Williamson
More broadly it will need to set out its approach to the science and technology framework for DSIT set out by the previous government in 2023 with its 10 priority areas Will this be revised? If so they need to set measurable targets and key outcomes in the priority areas. The government will also need to take a clear view on the key technologies we should be assisting in developing and commercialising
Then there are the pre existing financial commitments in the science and technology field. The Chancellor has said she will be checking all the previous government’s commitments for affordability. Which of the previous Government’s financial commitments will she confirm? For instance
The £7.4 million upskilling fund pilot to help SMEs develop AI skills.
Investing up to £100 million in the Alan Turing Institute over the next five years (up from £50 million)
The £100 million investment by the British Business Bank into ICG,in respect of the Long-term Investment for Technology and Science (LIFTS) initiative
The £1.1 billion funding for 65 Centres for Doctoral Training (CDTs) through the Engineering and Physical Sciences Research Council (EPSRC), covering key technologies like AI and engineering biology.
As regards the bills in the Kings Speech I look forward to seeing the details but the Digital Information and Smart Data bill does seem to be heading in the right direction in the areas being reinstated. The retention and enhancement of public trust in data use and sharing is the overriding need so that the potential of data can be unleashed through better trusted sharing of data. It is really important that we do more to educate the public about how and where our data is used and what powers individuals have to find out this information
I hope other than a few clarifications, especially in the research area, and in terms of the constitution of the ICO we are not going exhume some of the worst areas of the old DPDI bill and we have ditched the idea of a Brexit EU divergence Dividend by the watering down of so many data subject rights.
Will the Government give a firm commitment to safeguard our data adequacy with the EU? Will the bill introduce the promised ban on the creation of sexually explicit deepfakes? |
I also hope that the Government will confirm that the intent of the reinstated Digital Verification provisions is not compulsory national Digital ID but the creation of a market in digital ID providers that give choice to the citizen.
Given that LinesearchbeforeUdig, or LSBUD is claimed to already achieve the aims of NAUR, to be more widely used than the National Underground Assert Register NUAR and be more cost-effective, I hope also that Ministers will meet LSBUD and provide us all with much greater clarity around these proposals.
I hope that we can include other positive spects of the late unlamented DPDI Bill in the bill: More action on online fraud, digital identity theft, deepfakes in elections Misinformation and disinformation, misogyny as a hate crime, there is quite a list of possibilities. Together with new models of personal data control which were advocated as long ago as 2017 with the Hall Pesenti review, especially through new data communities and institutions and an enhanced ability to exercise our right to data portability, especially in real-time and more regulatory oversight over use of biometrics and biometric technologies.
I of course welcome the pledge to give coroners more powers to access information held by technology companies after a child’s death AND to banning the creation of sexually explicit deepfakes.
As regards the Cyber Security and Resilience Bill, events of recent days have made it clear we are not just talking about threats from bad actors. It reminds us how dependent we are on just a few overly dominant major tech companies. With Microsoft and AWS enjoying a combined UK market share of around 70-90%, according to the Competition and Markets Authority’s own research, the lack of competition presents a serious concerns for our nation's security and resilience. There needs to to be a rethink on critical national infrastructure such as cloud services and business software which are now essential public utilities and also how we are wholesale replacing reliable analogue communication with digital systems without backup.
In the bill I hope will we see the long awaited amendment of the Computer Misuse Act to include a statutory public interest defence, as called for by Cyber Up, to allow white hat research into computer systems as the Vallance report recommended. The rules for computer evidence must be changed too. We must have no more Horizon scandals!
Lords Committee Highly Critical of Office for Students
Shortly before parliament dissolved for the General Election the Housse of Lords debated the Report of the Industry and Regulators Committee Must do better: the Office for Students and the looming crisis facing higher education
https://publications.parliament.uk/pa/ld5803/ldselect/ldindreg/246/24602.htm
This is is what I said:
I declare an interest as chair of the council of Queen Mary University of London. I thank the noble Baroness, Lady Taylor, for her comprehensive and very fair introduction to our report. I thank her too for her excellent marshalling of our committee, with the noble Lord, Lord Hollick, and I add my thanks to our clerks and our special adviser during the inquiry.
I will speak in general terms rather than specifically about my own university. In higher education, there have been challenges aplenty to keep vice-chancellors and governing bodies awake at night: coming through the pandemic, industrial relations, cost of living rises for our students, pensions and research funding, to name but a few. But above all there are the ever-eroding unit of resource for domestic students, which was highlighted extremely effectively by the noble Lord, Lord Johnson of Marylebone, on the “Today” programme last week, and the Government’s continual policy interventions, including, above all, their seeming determination to reduce overseas student numbers.
In the face of this, I have to keep reminding myself that in 2021-22 Queen Mary University delivered a total economic benefit to the UK economy of £4.4 billion. For every pound we spent in 2021-22, we generated £7 of economic benefit. Universities are some of our great national assets. They not only are intellectual powerhouses for learning, education and social mobility, making a huge contribution to their local communities, but are inextricably linked to our national prospects for innovation and economic growth.
The committee’s report was very well received outside this House. Commentating in Wonkhe, the higher education blog—I do not know whether there are many readers of it around; I suspect there are—on the government and OfS responses to our report, its deputy editor noted:
“If you were expecting a seasonal mea culpa from either the regulator or the government … on any of these, it is safe to say that you will be disappointed”.
For him, the four standout aspects of our report were:
“the revelations about the place of students in the Office for Students … the criticism of the perceived closeness of the independent regulator to the government of the day … the school playground
level approach to the Designated Quality Body question … and the less splashy but deeply concerning suggestion that OfS didn’t really understand the financial problems the sector was facing”.
As regards the DQB question, which the committee explored in some depth, the current approach being taken by the OfS is extremely opaque. We clearly need a regulatory approach to quality to align with international standards. It is clear that the quickest way to get the English system realigned with international good practice would be to reinstate the QAA—an internationally recognised agency. Most of us cannot understand what seems to be the implacable hostility of the OfS to the QAA.
It is notable that the OfS, perhaps stung by the committee’s report, has now belatedly woken up to the fragility of the sector’s financial model and the fact that the future of the overseas student intake is central to financial underpinning. In its 2023 report on the financial sustainability of HE providers, the OfS confirmed that the
“overreliance on international student recruitment is a material risk for many types of providers where a sudden decline or interruption to international fees could trigger sustainability concerns”
and
“result in some providers having to make significant changes to their operating model or face a material risk of closure”.
Advice that they need to change their funding model and diversify their revenue streams is not particularly helpful, given the options available.
The Migration Advisory Committee’s Rapid Review of the Graduate Route, published last week—which recommends retaining the graduate visa on its current terms and reports that the graduate route is achieving the objectives set for it by the Government, finding
“no evidence of any significant abuse”—
is therefore of crucial importance. There is absolutely no doubt about the importance of the work study visa to the sector and the broader UK economy. In answer to the question from the noble Lord, Lord Parekh, we want it, and I hope that the OfS will play its part in trying to persuade the Government to retain it.
The Wonkhe blog also asks the fundamental question about the Government’s response regarding the regulatory burden on higher education. I hope the Minister can tell us: do the Government think it necessary and acceptable to keep ratcheting up regulation on universities? We are going in the wrong direction. Additional resource is required to monitor and provide returns in a whole variety of areas, such as the new freedom of speech requirements.
With the extraordinary contribution that universities make to society, communities and the economy as a whole, will university regulation benefit from the proposals set out in Smarter Regulation: Delivering a Regulatory Environment for Innovation, Investment and Growth, the Government’s recent White Paper? We will discuss this in a future debate on the response to our subsequent report, Who Watches the Watchdogs? For instance, the White Paper proposes the adoption of
“a culture of world-class service”
in how regulators undertake their day-to-day activities, and the adoption by all government departments of the
“10 principles of smarter regulation”.
It says:
“All government departmental annual reports must also set out the total costs and benefits of each significant regulation introduced that year”,
and says that the Government will
“strengthen the role of the Regulatory Policy Committee and the Better Regulation Framework, improve the assessment and scrutiny of the costs of regulation, and encourage non-regulatory alternatives”.
It says:
“The government will launch a Regulators Council to improve strategic dialogue between regulators and government, alongside monitoring the effectiveness of policy and strategic guidance issued”.
Finally, it says that
“it is up to the government to better assess its regulatory agenda, to try to understand the cost of its regulation on business and society”.
What is not to like, in the context of higher education regulation? Will all this be applied to the work of the OfS?
That all said, I welcome some of the way in which the OfS, if not the Government, has responded to our report. There is an air approaching contrition, in particular regarding engagement with both students and higher education providers. I welcome the OfS reviewing its approach to student engagement, empowering the student panel to raise issues that are important to students and increasing engagement with universities and colleges to improve sector relations
As regards the Government, a dialling down of their rhetoric continually undermining higher education, a pledge to ration ministerial directions given to the OfS, and putting university finances on a more sustainable, long-term footing would be welcome.
It is clear that continued scrutiny and evaluation— I very much liked what the noble Baroness, Lady Taylor, had to say about post-scrutiny reporting—will be essential to ensure that both government and OfS actions after their responses effectively address the underlying issues raised in our report. Sad to say, I do not think that the sector is holding its breath in the meantime.
Data Protection and Digital Information Bill lost in wash up-Hurray!
Having worked hard since the very first version of this bill (which was introduced into the Commons two years ago) to demonstrate its fundamental flaws, I and many other members of the Lords, and campaigning organisations, were delighted that the government failed to get it pushed through on so-called wash up when the General Election was called in May.
This is great news for citizens and for businesses too which would have faced the complications and expense of different compliance regimes in the EU and UK.
This is an edited version of what I said on Second Reading back in December 2023.
The Minister will have heard the concerns expressed throughout the House—not a single speaker failed to express concerns about the contents of the Bill. The retention and enhancement of public trust in data use and sharing is of key importance, but so much of the Bill seems almost entirely motivated by the Government’s desire to be divergent from the EU to get some kind of Brexit dividend.
As we have heard from all around the House, the Bill dilutes where it should strengthen the rights of data subjects. We can then all agree on the benefits of data sharing without the risks involved. The Equality and Human Rights Commission is clearly of that view, alongside numerous others, such as the Ada Lovelace Institute and as many as 26 privacy advocacy groups. Even on the Government’s own estimates, the Bill will have a minimal positive impact on compliance costs—in fact, it will simply lead to companies doing business in Europe having to comply with two sets of regulations.
I will be specific. I will go through a number of areas where I believe those rights are being diluted. The amended and more subjective definition of “personal data” will narrow the scope of what is considered personal data. Schedule 1 sets out a new annexe to the GDPR, with the types of processing activities that the Government have determined have a recognised legitimate interest and will not require a legitimate interest human rights balancing test to be carried out. Future Secretaries of State can amend or add to this list of recognised legitimate interests through secondary legislation. As a result it will become easier for political parties to target children as young as 14 during election campaigns, even though they cannot vote until they are 16 or 18, depending on the jurisdiction.
The Bill will change the threshold for refusing a subject access request, which will widen the grounds on which an organisation could refuse requests. There are existing difficulties of making those subject access requests. Clause 12, added on Report in the Commons, further tips power away from the individual’s ability to access data.
There are also changes to the automated decision-making provisions under Article 22 of the GDPR. The Bill replaces Article 22 with articles that reduce human review of automated decision-making. Article 22 should in fact be strengthened so that it applies to partly automated processing as well, and it should give rights to people affected by an automated decision, not just those who provide data. This should be the case especially in the workplace. A decision about you may be determined by data about other people whom you may never have met.
The Bill amends the circumstances in which personal datasets can be reused for research purposes. New clarifying guidance would have been sufficient, but for-profit commercial research is now included. As we discussed in debates on the then Online Safety Bill, the Bill does nothing where it really matters: on public interest researcher access.
The Bill moves away from UK GDPR requirements for mandatory data protection officers, and it also removes the requirement for data protection impact assessments. All this simply sets up a potential dual compliance system with less assurance—with what benefit? Under the new Bill, a controller or processor will be exempt from the duty to keep records, unless they are carrying out high-risk processing activities. But how effective will this be? One of the main ways of demonstrating compliance with GDPR is to have a record of processing activities.
There are also changes to the Information Commissioner’s role. We are all concerned about whether the creation of a new board will enable the ICO to maintain its current level of independence for data adequacy purposes. This is so important, as the noble Baroness, Lady Young, and my noble friend Lord McNally pointed out.
As regards intragroup transfers, there is concern from the National Aids Trust that Clause 5, permitting the intragroup transmission of personal health data
“where that is necessary for … administrative purposes”,
could mean that HIV/AIDS status is inadequately protected in workplace settings.
Schedule 5 to the Bill amends Chapter 5 of the UK GDPR to reform the UK’s regime for international transfers, with potential adverse consequences for business. There are the dangers of adopting too low standards internationally. This clearly has the potential to provide less protection for data subjects than the current test.
In Clause 17, the Bill removes a key enabler of collective interests, consultation with those affected by data and processing during the data protection risk assessment process, and it fails to provide alternative opportunities. Then there is the removal of the legal obligation to appoint a representative. This risks data breaches not being reported, takes away a channel of communication used by the ICO to facilitate its investigations, and increases the frustration of UK businesses in dealing with overseas companies that come to the UK market underprepared to comply with the UK GDPR.
Given that catalogue, it is hardly surprising that so many noble Lords have raised the issue of data adequacy. If I read out the list of all the noble Lords who have mentioned it, I would probably mention almost every single speaker in this debate. It is clear that the Bill significantly lowers data protection standards in the UK, as compared with the EU. On these Benches, our view is that this will undermine the basis of the UK’s EU data adequacy. The essential equivalence between the UK and the EU regimes has been critical to business continuity following Brexit. The Government’s own impact assessment acknowledges that, as the UK diverges from the EU GDPR, the risk of the EU revoking its adequacy decisions will increase. So I very much hope that the Minister, in response to all the questions he has been asked about data adequacy, has some pretty good answers, because there is certainly a considerable degree of concern around the House about the future of data adequacy.
In addition, there are aspects of the Bill that are just plain wrong. The Government need to deliver in full on their commitments to bereaved families made during the passage of what became the Online Safety Act, regarding access to their children’s data, in insisting that this is extended to all deaths of children. I very much hope that the Minister will harden up on his assurances at the end of the debate.
Noble Lords, questioned the abolition of the Surveillance Camera Commissioner, and the diminution of the duties relating to biometric data. Society is witnessing an unprecedented acceleration in the capability and reach of surveillance technologies, particularly live facial recognition, and we need the commissioner and Surveillance Camera Code of Practice in place. As the Ada Lovelace Institute says in its report Countermeasures, we need new and more comprehensive legislation on the use of biometrics, and the Equality and Human Rights Commission agrees with that too.
As regards the unrestrained financial powers, inserted at Commons Report stage, Sir Stephen Timms MP, chair of the DWP Select Committee, very rightly expressed strong concerns about this. These powers are entirely disproportionate and we will be strongly opposing them.
Then we have the new national security certificates and designation notices. These would give the Home Secretary great and unaccountable powers to authorise the police to violate our privacy rights, through the use of national security certificates and designation notices, without challenge. The Government have failed to explain why they believe these clauses are necessary to safeguard national security.
There is a whole series of missed opportunities during the course of the Bill. The Bill was an opportunity to create ethical, transparent and safe standards for AI systems. A number of noble Lords all said that this is a wasted opportunity to create measures adequate to an era of ubiquitous use of data through AI systems. The noble Baroness, Lady Kidron, in particular talked about this in relation to children, generative AI and educational technology. It is so important in the public sector as well.
The EU has just agreed in principle to a new AI Act. We are miles behind the curve. Then, of course, we have the new identification verification framework. The UK has chosen not to allow private sector digital ID systems to be used for access to public servives. Perhaps the Government could explain why that is the case.
There are a number of other areas, such as new models of personal data control, which were advocated as long ago as 2017, with the Hall-Pesenti review- data instritutions, comunities and truits. Why are the Government not being more imaginative in that sense? There is also the avoidance of creating a new offence of identity theft. That seems to be a great missed opportunity in this Bill.
There is the question of holding AI system providers to be legally accountable for the generation of child sexual abuse material online by using their datasets. . Why are the Government not taking the opportunity to correct the the case of ICO v Experian ?
In the face of the need to do more to protect citizens’ rights, this Bill is a dangerous distraction. It waters down rights, it is a huge risk to data adequacy, it is wrong in many areas and it is a great missed opportunity in many others. We on these Benches will oppose a Bill which appears to have very few friends around the House. We want to amend a great many of the provisions of the Bill and we want to scrutinise many other aspects of it where the amendments came through at a very late stage. I am afraid the Government should expect this Bill to have a pretty rough passage.
We Need a New Offence of Digital ID Theft
As part of the debates on the Data Protection Bill I recently advocated for a new Digital ID theft offence . This is what i said.
It strikes me as rather extraordinary that we do not have an identity theft offence. This is the Metropolitan Police guidance for the public:
“Your identity is one of your most valuable assets. If your identity is stolen, you can lose money and may find it difficult to get loans, credit cards or a mortgage. Your name, address and date of birth provide enough information to create another ‘you’”.
It could not be clearer. It goes on:
“An identity thief can use a number of methods to find out your personal information and will then use it to open bank accounts, take out credit cards and apply for state benefits in your name”.
It then talks about the signs that you should look out for, saying:
“There are a number of signs to look out for that may mean you are or may become a victim of identity theft … If you think you are a victim of identity theft or fraud, act quickly to ensure you are not liable for any financial losses … Contact CIFAS (the UK’s Fraud Prevention Service) to apply for protective registration”.
However, there is no criminal offence.
Interestingly enough, I mentioned this to the noble Baroness, Lady Morgan; Back in October 2022, her committee—the Fraud Act 2006 and Digital Fraud Committee—produced a really good report, Fighting Fraud: Breaking the Chain, which said:
“Identity theft is often a predicate action to the criminal offence of fraud, as well as other offences including organised crime and terrorism, but it is not a criminal offence. Cifas datashows that cases of identity fraud increased by 22% in 2021, accounting for 63% of all cases recorded to Cifas’ National Fraud Database”.
It goes on to talk about identity theft to some good effect but states:
“In February 2022, the Government confirmed that there were no plans to introduce a new criminal offence of identity theft as ‘existing legislation is in place to protect people’s personal data and prosecute those that commit crimes enabled by identity theft’”.
I do not think the committee agreed with that at all. It said:
“The Government should consult on the introduction of legislation to create a specific criminal offence of identity theft. Alternatively, the Sentencing Council should consider including identity theft as a serious aggravating factor in cases of fraud”.
The Government are certainly at odds with the Select Committee chaired by the noble Baroness, Lady Morgan. I am indebted to a creative performer called Bennett Arron, who raised this with me some years ago. He related with some pain how he took months to get back his digital identity. He said: “I eventually, on my own, tracked down the thief and gave his name and address to the police. Nothing was done. One of the reasons the police did nothing was because they didn’t know how to charge him with what he had done to me”. That is not a good state of affairs. Then we heard from Paul Davis, the head of fraud prevention at TSB. The headline of the piece in the Sunday Times was: “I’m head of fraud at a bank and my identity was still stolen”. He is top dog in this area, and he has been the subject of identity theft.
This seems an extraordinary situation, whereby the Government are sitting on their hands. There is a clear issue with identity theft, yet they are refusing—they have gone into print, in response to the committee chaired by the noble Baroness, Lady Morgan—and saying, “No, no, we don’t need anything like that; everything is absolutely fine”. I hope that the Minister can give a better answer this time around.
Lords Debate Report on AI in Weapon Systems
Recently the House of Lords Debated the Report of the AI in Weapon Systems Committee Proceed with Caution.
This is an edited version of what I said
Autonomous weapon systems present some of the most emotive and high-risk challenges posed by AI. We have heard a very interesting rehearsal of some of the issues surrounding use and possible benefits, but particularly the risks. I believe that the increasing use of drones in particular, potentially autonomously, in conflicts such as Libya, Syria and Ukraine and now by Iran and Israel, together with AI targeting systems such as Lavender, highlights the urgency of addressing the governance of weapon systems.
The implications of autonomous weapons systems—AWS—are far-reaching. There are serious risks to consider, such as escalation and proliferation of conflict, accountability and lack of accountability for actions,
and cybersecurity vulnerabilities. There is the lack of empathy and kindness qualities that humans are capable of in making military decisions. There is misinformation and disinformation, which is a new kind of warfare.
Professor Stuart Russell, in his Reith lecture on this subject in 2021, painted a stark picture of the risks posed by scalable autonomous weapons capable of destruction on a mass scale. This chilling scenario underlines the urgency with which we must approach the regulation of AWS. The UK military sees AI as a priority for the future, with plans to integrate “boots and bots” to quote a senior military officer.
The UK integrated review of 2021 made lofty commitments to ethical AI development. Despite this and the near global consensus on the need to regulate AWS, the UK has not yet endorsed limitations on their use. The UK’s defence AI strategy and its associated policy statement, Ambitious, Safe, Responsible, acknowledged the line that should not be crossed regarding machines making combat decisions but lacked detail on where this line is drawn, raising ethical, legal and indeed moral concerns.
As we explored this complex landscape as a committee—and it was quite a journey for many of us—we found that, while the term AWS is frequently used, its definition is elusive. The inconsistency in how we define and understand AWS has significant implications for the development and governance of these technologies. However, the committee demonstrated that a working definition is possible, distinguishing between fully and partially autonomous systems. This is clearly still resisted by the Government, as their response has shown.
The current lack of definition allows for the assertion that the UK neither possesses nor intends to develop fully autonomous systems, but the deployment of autonomous systems raises questions about accountability, especially in relation to international humanitarian law. The Government emphasise the sufficiency of existing international humanitarian law while a human element in weapon deployment is retained. The Government have consistently stated that UK forces do not use systems that deploy lethal force without human involvement, and I welcome that.
Despite the UK’s reluctance to limit AWS, the UN and other states advocate for specific regulation. The UN Secretary-General, António Guterres, has called autonomous weapons with life-and-death decision-making powers “politically unacceptable, morally repugnant” and deserving of prohibition, yet an international agreement on limitation remains elusive.
In our view, the rapid development and deployment of AWS necessitates regulatory frameworks that address the myriad of challenges posed by these technologies. The relationship between our own military and the private sector makes it even more important that we address the challenges posed by these technologies and ensure compliance with international law to maintain ethical standards and human oversight. I share the optimism of the noble Lord, Lord Holmes, that this is both possible and necessary.
Human rights organisations have urged the UK to lead in establishing new international law on autonomous weapon systems to address the current deadlock in conventional weapons conventions, and we should do so. There is a clear need for the UK to play an active role in shaping the nature of future military engagement.
A historic moment arrived last November with the UN’s first resolution on autonomous weapons, affirming the application of international law to these systems and setting the stage for further discussion at the UN General Assembly. The UK showed support for the UN resolution that begins consultations on these systems, which I very much welcome. The Government have committed also to explicitly ensure human control at all stages of an AWS’s life cycle. It is essential to have human control over the deployment of the system, to ensure both human moral agency and compliance with international humanitarian law.
However, the Government still have a number of questions to answer. Will they respond positively to the call by the UN Secretary-General and the International Committee of the Red Cross that a legally binding instrument be negotiated by states by 2026? How do the Government intend to engage at the Austrian Government’s conference “Humanity at the Crossroads”, which is taking place in Vienna at the end of this month? What is the Government’s assessment of the implications of the use of AI targeting systems under international humanitarian law? Can the Government clarify how new international law on AWS would be a threat to our defence interests? What factors are preventing the Government adopting a definition of AWS, as the noble Lord, Lord Lisvane, asked? What steps are being taken to ensure meaningful human involvement throughout the life cycle of AI-enabled military systems? Finally, will the Government continue discussions at the Convention on Certain Conventional Weapons, and continue to build a common understanding of autonomous weapon systems and elements of the constraints that should be placed on them?
The committee rightly warns that time is short for us to tackle the issues surrounding AWS. I hope the Government will pay close and urgent attention to its recommendations.
Lord Holmes Private Members bill a "stake in the ground" says Lord C-J
Lord Holmes of Richmond recently introduced his Private Members Bill -The Artificial Intelligence (Regulation) Bill.
This may not go as far in regulating AI as many want to see but it is a good start. This what Lord Holmes says about it on his own website
https://lordchrisholmes.com/artificial-intelligence-regulation-bill/
and this is what I said at its second reading recently
My Lords, I congratulate the noble Lord, Lord Holmes, on his inspiring introduction and on stimulating such an extraordinarily good and interesting debate.
The excellent House of Lords Library guide to the Bill warns us early on:
“The bill would represent a departure from the UK government’s current approach to the regulation of AI”.
Given the timidity of the Government’s pro-innovation AI White Paper and their response, I would have thought that was very much a “#StepInTheRightDirection”, as the noble Lord, Lord Holmes, might say.
There is clearly a fair wind around the House for the Bill, and I very much hope it progresses and we see the Government adopt it, although I am somewhat pessimistic about that. As we have heard in the debate, there are so many areas where AI is and can potentially be hugely beneficial. However, as many noble Lords have emphasised, it also carries risks, not just of the existential kind, which the Bletchley Park summit seemed to address, but others mentioned by noble Lords today, such as misinformation, disinformation, child sexual abuse, and so on, as well as the whole area of competition—the issue of the power and the asymmetry of these big tech AI systems and the danger of regulatory capture.
It is disappointing that, after a long gestation of national AI policy-making, which started so well back in 2017 with the Hall-Pesenti review, contributed to by our own House of Lords Artificial Intelligence Committee, the Government have ended up by producing a minimalist approach to AI regulation. I liked the phrase used by the noble Lord, Lord Empey, “lost momentum”, because it certainly feels like that after this period of time.
The UK’s National AI Strategy, a 10-year plan for UK investment in and support of AI, was published in September 2021 and accepted that in the UK we needed to prepare for artificial general intelligence. We needed to establish public trust and trustworthy AI, so often mentioned by noble Lords today. The Government had to set an example in their use of AI and to adopt international standards for AI development and use. So far, so good. Then, in the subsequent AI policy paper, AI Action Plan, published in 2022, the Government set out their emerging proposals for regulating AI, in which they committed to develop
“a pro-innovation national position on governing and regulating AI”,
to be set out in a subsequent governance White Paper. The Government proposed several early cross-sectoral and overarching principles that built on the OECD principles on artificial intelligence: ensuring safety, security, transparency, fairness, accountability and the ability to obtain redress.
Again, that is all good, but the subsequent AI governance White Paper in 2023 opted for a “context-specific approach” that distributes responsibility for embedding ethical principles into the regulation of AI systems across several UK sector regulators without giving them any new regulatory powers. I thought the analysis of this by the noble Lord, Lord Young, was interesting. There seemed to be no appreciation that there were gaps between regulators. That approach was confirmed this February in the response to the White Paper consultation.
Although there is an intention to set up a central body of some kind, there is no stated lead regulator, and the various regulators are expected to interpret and apply the principles in their individual sectors in the expectation that they will somehow join the dots between them. There is no recognition that the different forms of AI are technologies that need a comprehensive cross-sectoral approach to ensure that they are transparent,
explainable, accurate and free of bias, whether they are in an existing regulated or unregulated sector. As noble Lords have mentioned, discussing existential risk is one thing, but going on not to regulate is quite another.
Under the current Data Protection and Digital Information Bill, data subject rights regarding automated decision-making—in practice, by AI systems—are being watered down, while our creatives and the creative industries are up in arms about the lack of support from government in asserting their intellectual property rights in the face of the ingestion of their material by generative AI developers. It was a pleasure to hear what the noble Lord, Lord Freyberg, had to say on that.
For me, the cardinal rules are that business needs clarity, certainty and consistency in the regulatory system if it is to develop and adopt AI systems, and we need regulation to mitigate risk to ensure that we have public trust in AI technology. Regulation is not necessarily the enemy of innovation; it can be a stimulus. That is something that we need to take away from this discussion.
This is where the Bill of the noble Lord, Lord Holmes, is an important stake in the ground, as he has described. It provides for a central AI authority that has a duty of looking for gaps in regulation; it sets out extremely well out the safety and ethical principles to be followed; it provides for regulatory sandboxes, which we should not forget are an innovation invented in the UK; and it provides for AI responsible officers and for public engagement. Importantly, it builds in a duty of transparency regarding data and IP-protected material where they are used for training purposes, and for labelling AI-generated material, as the noble Baroness, Lady Stowell, and her committee have advocated. By itself, that would be a major step forward, so, as the noble Lord knows, we on these Benches wish the Bill very well, as do all those with an interest in protecting intellectual property, as we heard the other day at the round table that he convened.
However, in my view what is needed at the end of the day is the approach that the interim report of the Science, Innovation and Technology Committee recommended towards the end of last year in its inquiry into AI governance: a combination of risk-based cross-sectoral regulation and specific regulation in sectors such as financial services, applying to both developers and adopters, underpinned by common trustworthy standards of risk assessment, audit and monitoring. That should also provide recourse and redress, as the Ada Lovelace Institute, which has done so much work in the area, asserts.
That should include the private sector, where there is no effective regulator for the workplace, mentioned, and the public sector, where there is no central or local government compliance mechanism; no transparency yet in the form of a public register of use of automated decision-making, despite the promised adoption of the algorithmic recording standard; and no recognition by the Government that explicit legislation and/or regulation for intrusive
AI technologies used in the public sector, such as live facial recognition and other biometric capture, is needed. Then, of course, we need to meet the IP challenge. We need to introduce personality rights to protect our artists, writers and performers. We need the labelling of AI-generated material alongside the kinds of transparency duties contained in the noble Lord’s Bill.
Then there is another challenge, which is more international. We have world-beating AI researchers and developers. How can we ensure that, despite differing regulatory regimes—for instance, between ourselves and the EU or the US—developers are able to commercialise their products on a global basis and adopters can have the necessary confidence that the AI product meets ethical standards?
The answer, in my view, lies in international agreement on common standards such as those of risk and impact assessment, testing, audit, ethical design for AI systems, and consumer assurance, which incorporate what have become common internationally accepted AI ethics. Having a harmonised approach to standards would help provide the certainty that business needs to develop and invest in the UK more readily, irrespective of the level of obligation to adopt them in different jurisdictions and the necessary public trust. In this respect, the UK has the opportunity to play a much more positive role with the Alan Turing Institute’s AI Standards Hub and the British Standards Institution. The OECD.AI group of experts is heavily involved in a project to find common ground between the various standards.
We need a combination of proportionate but effective regulation in the UK and the development of international standards, so, in the words of the noble Lord, Lord Holmes, why are we not legislating? His Bill is a really good start; let us build on it.
New Digital Markets Bill Must Not be Watered Down
The Digital Markets Competition and Consumer Bill had its Second Reading in the House of Lords on the 5th December 2023 and its 3rd Reading on the 26th March 2024 This is an edited version of what I said on each occasion
Second Reading
I thank the Minister for what I thought was a comprehensive introduction that really set the scene for the Bill. As my noble friend said, we very much welcome the Bill, broadly. It is an overdue offspring of the Furman review and, along with so many noble Lords around the House, he gave very cogent reasons, given the dominance that big tech has and the inadequate powers that our competition regulators have had to tackle them. It is absolutely clear around the House that there is great appetite for improving the Bill. I have knocked around this House for a few years, and I have never heard such a measure of agreement at Second Reading.
We seem to have repeated ourselves, but repetition is good. I am sure that in the Minister’s notebook he just has a list saying “agree, agree, agree” as we have gone through the Bill. I very much hope that he will follow the example that both he and the noble Lord, Lord Parkinson, demonstrated on the then Online Safety Bill and will engage across and around the Chamber with all those intervening today, so that we really can improve the Bill.
It is not just size that matters: we must consider behaviour, dominance, market failure and market power. We need to hold on to that. We need new, flexible pro-competition powers and the ability to act ex ante and on an interim basis—those are crucial powers for the CMA. As we have heard from all round the House, the digital landscape, whether it is app stores, cloud services or more, is dominated by the power of certain big tech companies, particularly in AI, with massive expenditure on compute power, advanced semiconductors, large datasets and the scarce technology skills forming a major barrier to entry where the development of generative AI is concerned. We can already see the future coming towards us.
In that context, I very much welcome Ofcom’s decision to refer the hyperscalers in cloud services for an investigation by the CMA. The CMA and the DMU have the capability to deliver the Bill’s aims.the It must have the ability to implement the new legislative powers. Unlike some other commentators, we believe, as my noble friend said, that the CMA played a positively useful role in the Activision Blizzard-Microsoft merger. It is crucial that the CMA is independent of government. All around the House, there was comment about the new powers of the Secretary of State in terms of guidance. The accountability to Parliament will also be crucial, and that was again a theme that came forward. We heard about the Joint Committee proposals made by both the committee of the noble Baroness, Lady Stowell, and the Joint Committee on the Online Safety Bill.
We need to ensure that that scrutiny is there and, as the Communications and Digital Committee also said, that the DMU is well resourced and communicates its priorities, work programmes and decisions regularly to external stakeholders and Parliament.
The common theme across this debate—to mention individual noble Lords, I would have to mention almost every speaker—has been that the Bill must not be watered down. In many ways, that means going back to the original form of the Bill before it hit Report in the Commons. We certainly very much support that approach, whether it is to do with the merits approach to penalties, the explicit introduction of proportionality or the question of deleting the indispensability test in the countervailing benefits provisions. We believe that, quite apart from coming back on the amendments from Report, the Bill could be further strengthened in a number of respects.
In the light of the recent Open Markets Institute report, we should be asking whether we are going far enough in limiting the power of big tech. In particular, as regards the countervailing benefits exemption, as my noble friend said, using the argument of countervailing benefits—even if we went back to the definition from Report—must not be used by big tech as a major loophole to avoid regulatory action. It is clear that many noble Lords believe, especially in the light of those amendments, that the current countervailing benefits exemption provides SMS firms with too much room to evade conduct requirements.
The key thing that unites us is the fact that, even though we must act in consumers’ interests, this is not about short-term consumer welfare but longer-term consumers’ interests; a number of noble Lords from across the House have made that really important distinction.
We believe that there should be pre-notification if a platform intends to rely on this exemption. The scope of the exemption should also be significantly curtailed to prevent its abuse, in particular by providing an exhaustive list of the types of countervailing benefits that SMS firms are able to claim. We would go further in limiting the way in which the exemption operates.
On strategic market status, one of the main strengths of the Bill is its flexible approach. However, the current five-year period does not account for dynamic digital markets that will not have evidence of the position in the market in five years’ time. We believe that the Bill should be amended so that substantial and entrenched market power is mainly based on past data rather than a forward-looking assessment, and that the latter is restricted to a two-year assessment period. The consultation aspect of this was also raised; there should be much greater rights on the consultation of businesses that are not of strategic market status under the Bill.
A number of noble Lords recognised the need for speed. It is not just a question of making sure that the CMA has the necessary powers; it must be able to move quickly. We believe that the CMA should be given the legal power to secure injunctions under the High Court timetable, enabling it to stop anti-competitive activities in days. This would be in addition to the CMA’s current powers.
We have heard from across the House about the final offer mechanism affecting the news media. We believe that a straightforward levy on big tech platforms, redistributed to smaller journalism enterprises, would be a far more equitable approach. However we need to consider in the context of the Bill the adoption by the CMA of the equivalent to Ofcom’s duty in the Communications Act 2003
“to further the interests of citizens”,
so that it must consider the importance of an informed democracy and a plural media when considering its remedies.
The Bill needs to make it clear that platforms need to pay properly and fairly for content, on benchmarked terms and with reference to value for end-users. Indeed, we believe that they must seek permission for the content that they use. As we heard from a number of noble Lords, that is becoming particularly important as regards the large language models currently being developed.
We also believe it is crucial that smaller publishers are not frozen out or left with small change while the highly profitable large publishers scoop the pool. I hope that we will deal with the Daily Telegraph ownership question and the mergers regime in the Enterprise Act as we go forward into Committee, to make sure that the accumulation of social media platforms is assessed beyond the purely economic perspective. The Enterprise Act powers should be updated to allow the Secretary of State to issue a public interest notice seeking Ofcom’s advice on digital media mergers, as well as newspapers, and at the lower thresholds proposed by this Bill.
There were a number of questions related to leveraging. We want to make sure that we have the right approach to that. The Bill does not seem to be drafted properly in allowing the CMA to prevent SMS firms using their dominance in designated activities to increase their power in non-designated activities. We want to kick the tyres on that.
Of course, there are a great many consumer protection issues here, which a number of noble Lords raised. They include fake reviews and the need for collective action. It is important that we allow collective action not just on competition rights but further, through consumer claims, data abuse claims and so on. We should cap the costs for claimants in the Competition Appeal Tribunal.These issues also include misleading packaging.
Nearly every speaker mentioned subscriptions. I do not think that I need to point out to the Minister the sheer unanimity on this issue. We need to get this right because there is clearly support across the House for making sure that we get the provisions right while protecting the income of charities.
There is a whole host of other issues that we will no doubt discuss in Committee: mid-contract price rises, drip pricing, ticket touting, online scams and reforming ADR. We want to see this Bill and the new competition and consumer powers make a real difference. However, we believe that we can do this only with some key changes being made to the Bill, which are clearly common ground between us all, as we have debated the Bill today. We look forward to the Committee proceedings next year—I can say that now—which will, I hope, be very productive, if both Ministers will it so.
Third Reading
I reiterate the welcome that we on these Benches gave to the Bill at Second Reading. We believe it is vital to tackle the dominance of big tech and to enhance the powers of our competition regulators to tackle it, in particular through the new flexible pro-competition powers and the ability to act ex ante and on an interim basis.
We were of the view, and still are, that the Bill needs strengthening in a number of respects. We have been particularly concerned about the countervailing benefits exemption under Clause 29. This must not be used by big tech as a major loophole to avoid regulatory action. A number of other aspects were inserted into the Bill on Report in the Commons about appeals standards and proportionality. During the passage of the Bill, we added a fourth amendment to ensure that the Secretary of State’s power to approve CMA guidance will not unduly delay the regime coming into effect.
As the noble Baroness, Lady Stowell, said, we are already seeing big tech take an aggressive approach to the EU Digital Markets Act. We therefore believe the Bill needs to be more robust in this respect. In this light, it is essential to retain the four key amendments passed on Report and that they are not reversed through ping-pong when the Bill returns to the Commons.
I thank both Ministers and the Bill team. They have shown great flexibility in a number of other areas, such as online trading standards powers, fake reviews, drip pricing, litigation, funding, cooling-off periods, subscriptions and, above all, press ownership, as we have seen today. They have been assiduous in their correspondence throughout the passage of the Bill, and I thank them very much for that, but in the crucial area of digital markets we have seen no signs of movement. This is regrettable and gives the impression that the Government are unwilling to move because of pressure from big tech. If the Government want to dispel that impression, they should agree with these amendments, which passed with such strong cross-party support on Report.
In closing, I thank a number of outside organisations that have been so helpful during the passage of the Bill—in particular, the Coalition for App Fairness, the Public Interest News Foundation, Which?, Preiskel & Co, Foxglove, the Open Markets Institute and the News Media Association. I also thank Sarah Pughe and Mohamed-Ali Souidi in our own Whips’ Office.
Given the coalition of interest that has been steadily building across the House during the debates on the Online Safety Bill and now this Bill, I thank all noble Lords on other Benches who have made common cause and, consequently, had such a positive impact on the passage of this Bill. As with the Online Safety Act, this has been a real collaborative effort in a very complex area.