Category: Tech News

  • The Cultural Impact of AI Generated Content: Part 2

    Stephanie Kirmer

    What can we do about the increasingly sophisticated AI generated content in our lives?

    Photo by Meszárcsek Gergely on Unsplash

    In my prior column, I established how AI generated content is expanding online, and described scenarios to illustrate why it’s occurring. (Please read that before you go on here!) Let’s move on now to talking about what the impact is, and what possibilities the future might hold.

    Social and Creative Creatures

    Human beings are social creatures, and visual ones as well. We learn about our world through images and language, and we use visual inputs to shape how we think and understand concepts. We are shaped by our surroundings, whether we want to be or not.

    Accordingly, no matter how much we are consciously aware of the existence of AI generated content in our own ecosystems of media consumption, our subconscious response and reaction to that content will not be fully within our control. As the truism goes, everyone thinks they’re immune to advertising — they’re too smart to be led by the nose by some ad executive. But advertising continues! Why? Because it works. It inclines people to make purchasing choices that they otherwise wouldn’t have, whether just from increasing brand visibility, to appealing to emotion, or any other advertising technique.

    AI-generated content may end up being similar, albeit in a less controlled way. We’re all inclined to believe we’re not being fooled by some bot with an LLM generating text in a chat box, but in subtle or overt ways, we’re being affected by the continued exposure. As much as it may be alarming that advertising really does work on us, consider that with advertising the subconscious or subtle effects are being designed and intentionally driven by ad creators. In the case of generative AI, a great deal of what goes into creating the content, no matter what its purpose, is based on an algorithm using historical information to choose the features most likely to appeal, based on its training, and human actors are less in control of what that model generates.

    I mean to say that the results of generative AI routinely surprise us, because we’re not that well attuned to what our history really says, and we often don’t think of edge cases or interpretations of prompts we write. The patterns that AI is uncovering in the data are sometimes completely invisible to human beings, and we can’t control how these patterns influence the output. As a result, our thinking and understanding are being influenced by models that we don’t completely understand and can’t always control.

    Critical Thinking

    Beyond that, as I’ve mentioned, public critical thinking and critical media consumption skills are struggling to keep pace with AI generated content, to give us the ability to be as discerning and thoughtful as the situation demands. Similarly to the development of Photoshop, we need to adapt, but it’s unclear whether we have the ability to do so.

    We are all learning tell-tale signs of AI generated content, such as certain visual clues in images, or phrasing choices in text. The average internet user today has learned a huge amount in just a few years about what AI generated content is and what it looks like. However, suppliers of the models used to create this content are trying to improve their performance to make such clues subtler, attempting to close the gap between obviously AI generated and obviously human produced media. We’re in a race with AI companies, to see whether they can make more sophisticated models faster than we can learn to spot their output.

    We’re in a race with AI companies, to see whether they can make more sophisticated models faster than we can learn to spot their output.

    In this race, it’s unclear if we will catch up, as people’s perceptions of patterns and aesthetic data have limitations. (If you’re skeptical, try your hand at detecting AI generated text: https://roft.io/) We can’t examine images down to the pixel level the way a model can. We can’t independently analyze word choices and frequencies throughout a document at a glance. We can and should build tools that help do this work for us, and there are some promising approaches for this, but when it’s just us facing an image, a video, or a paragraph, it’s just our eyes and brains versus the content. Can we win? Right now, we often don’t. People are fooled every day by AI-generated content, and for every piece that gets debunked or revealed, there must be many that slip past us unnoticed.

    One takeaway to keep in mind is that it’s not just a matter of “people need to be more discerning” — it’s not as simple as that, and if you don’t catch AI generated materials or deepfakes when they cross your path every time, it’s not all your fault. This is being made increasingly difficult on purpose.

    Plus, Bots!

    So, living in this reality, we have to cope with a disturbing fact. We can’t trust what we see, at least not in the way we have become accustomed to. In a lot of ways, however, this isn’t that new. As I described in my first part of this series, we kind of know, deep down, that photographs may be manipulated to change how we interpret them and how we perceive events. Hoaxes have been perpetuated with newspapers and radio since their invention as well. But it’s a little different because of the race — the hoaxes are coming fast and furious, always getting a little more sophisticated and a little harder to spot.

    We can’t trust what we see, at least not in the way we have become accustomed to.

    There’s also an additional layer of complexity in the fact that a large amount of the AI generated content we see, particularly on social media, is being created and posted by bots (or agents, in the new generative AI parlance), for engagement farming/clickbait/scams and other purposes as I discussed in part 1 of this series. Frequently we are quite a few steps disconnected from a person responsible for the content we’re seeing, who used models and automation as tools to produce it. This obfuscates the origins of the content, and can make it harder to infer the artificiality of the content by context clues. If, for example, a post or image seems too good (or weird) to be true, I might investigate the motives of the poster to help me figure out if I should be skeptical. Does the user have a credible history, or institutional affiliations that inspire trust? But what if the poster is a fake account, with an AI generated profile picture and fake name? It only adds to the challenge for a regular person to try and spot the artificiality and avoid a scam, deepfake, or fraud.

    As an aside, I also think there’s general harm from our continued exposure to unlabeled bot content. When we get more and more social media in front of us that is fake and the “users” are plausibly convincing bots, we can end up dehumanizing all social media engagement outside of people we know in analog life. People already struggle to humanize and empathize through computer screens, hence the longstanding problems with abuse and mistreatment online in comments sections, on social media threads, and so on. Is there a risk that people’s numbness to humanity online worsens, and degrades the way they respond to people and models/bots/computers?

    What Now?

    How do we as a society respond, to try and prevent being taken in by AI-generated fictions? There’s no amount of individual effort or “do your homework” that can necessarily get us out of this. The patterns and clues in AI-generated content may be undetectable to the human eye, and even undetectable to the person who built the model. Where you might normally do online searches to validate what you see or read, those searches are heavily populated with AI-generated content themselves, so they are increasingly no more trustworthy than anything else. We absolutely need photographs, videos, text, and music to learn about the world around us, as well as to connect with each other and understand the broader human experience. Even though this pool of material is becoming poisoned, we can’t quit using it.

    There are a number of possibilities for what I think might come next that could help with this dilemma.

    • AI declines in popularity or fails due to resource issues. There are a lot of factors that threaten the growth and expansion of generative AI commercially, and these are mostly not mutually exclusive. Generative AI very possibly could suffer some degree of collapse due to AI generated content infiltrating the training datasets. Economic and/or environmental challenges (insufficient power, natural resources, or capital for investment) could all slow down or hinder the expansion of AI generation systems. Even if these issues don’t affect the commercialization of generative AI, they could create barriers to the technology’s progressing further past the point of easy human detection.
    • Organic content becomes premium and gains new market appeal. If we are swarmed with AI generated content, that becomes cheap and low quality, but the scarcity of organic, human-produced content may drive a demand for it. In addition, there is a significant growth already in backlash against AI. When customers and consumers find AI generated material off-putting, companies will move to adapt. This aligns with some arguments that AI is in a bubble, and that the excessive hype will die down in time.
    • Technological work challenges the negative effects of AI. Detector models and algorithms will be necessary to differentiate organic and generated content where we can’t do it ourselves, and work is already going on in this direction. As generative AI grows in sophistication, making this necessary, a commercial and social market for these detector models may develop. These models need to become a lot more accurate than they are today for this to be possible — we don’t want to rely upon notably bad models like those being used to identify generative AI content in student essays in educational institutions today. But, a lot of work is being done in this space, so there’s reason for hope. (I have included a few research papers on these topics in the notes at the end of this article.)
    • Regulatory efforts expand and gain sophistication. Regulatory frameworks may develop sufficiently to be helpful in reining in the excesses and abuses generative AI enables. Establishing accountability and provenance for AI agents and bots would be a massively positive step. However, all this relies on the effectiveness of governments around the world, which is always uncertain. We know big tech companies are intent on fighting against regulatory obligations and have immense resources to do so.

    I think it very unlikely that generative AI will continue to gain sophistication at the rate seen in 2022–2023, unless a significantly different training methodology is developed. We are running short of organic training data, and throwing more data at the problem is showing diminishing returns, for exorbitant costs. I am concerned about the ubiquity of AI-generated content, but I (optimistically) don’t think these technologies are going to advance at more than a slow incremental rate going forward, for reasons I have written about before.

    This means our efforts to moderate the negative externalities of generative AI have a pretty clear target. While we continue to struggle with difficulty detecting AI-generated content, we have a chance to catch up if technologists and regulators put the effort in. I also think it is vital that we work to counteract the cynicism this AI “slop” inspires. I love machine learning, and I’m very glad to be a part of this field, but I’m also a sociologist and a citizen, and we need to take care of our communities and our world as well as pursuing technical progress.

    Read more of my work at www.stephaniekirmer.com.

    Further Reading

    https://dl.acm.org/doi/pdf/10.1145/3637528.3671463

    https://arxiv.org/pdf/2402.00045

    https://iris.uniroma1.it/bitstream/11573/1710645/1/Maiano_Human_2024.pdf

    https://dl.acm.org/doi/abs/10.1145/3658644.3670306

    https://www.cis.upenn.edu/~ccb/publications/miragenews.pdf


    The Cultural Impact of AI Generated Content: Part 2 was originally published in Towards Data Science on Medium, where people are continuing the conversation by highlighting and responding to this story.

    Originally appeared here:
    The Cultural Impact of AI Generated Content: Part 2

    Go Here to Read this Fast! The Cultural Impact of AI Generated Content: Part 2

  • What I’m Updating in My AI Ethics Class for 2025

    What I’m Updating in My AI Ethics Class for 2025

    Nathan Bos, Ph.D.

    What happened in 2024 that is new and significant in the world of AI ethics? The new technology developments have come in fast, but what has ethical or values implications that are going to matter long-term?

    I’ve been working on updates for my 2025 class on Values and Ethics in Artificial Intelligence. This course is part of the Johns Hopkins Education for Professionals program, part of the Master’s degree in Artificial Intelligence.

    Overview of my changes:

    I’m doing major updates on three topics based on 2024 developments, and a number of small updates, integrating other news and filling gaps in the course.

    Topic 1: LLM interpretability.

    Anthropic’s work in interpretability was a breakthrough in explainable AI (XAI). We will be discussing how this method can be used in practice, as well as implications for how we think about AI understanding.

    Topic 2: Human-Centered AI.

    Rapid AI development adds urgency to the question: How do we design AI to empower rather than replace human beings? I have added content throughout my course on this, including two new design exercises.

    Topic 3: AI Law and Governance.

    Major developments were the EU’s AI Act and the raft of California legislation, including laws targeting deep fakes, misinformation, intellectual property, medical communications and minor’s use of ‘addictive’ social media, among other. For class I developed some heuristics for evaluating AI legislation, such as studying definitions, and explain how legislation is only one piece of the solution to the AI governance puzzle.

    California Governor Gavin Newsome signing one of several new AI laws; public domain photo by the State of California

    Miscellaneous new material:

    I am integrating material from news stories into existing topics on copyright, risk, privacy, safety and social media/ smartphone harms.

    Topic 1: Generative AI Interpretability

    What’s new:

    Anthropic’s pathbreaking 2024 work on interpretability was a fascination of mine. They published a blog post here, and there is also a paper, and there was an interactive feature browser. Most tech-savvy readers should be able to get something out of the blog and paper, despite some technical content and a daunting paper title (‘Scaling Monosemanticity’).

    Below is a screenshot of one discovered feature, ‘syncophantic praise’. I like this one because of the psychological subtlety; it amazes me that they could separate this abstract concept from simple ‘flattery’, or ‘praise’.

    Graphic from the paper ‘Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet’.

    What’s important:

    Explainable AI: For my ethics class, this is most relevant to explainable AI (XAI), which is a key ingredient of human-centered design. The question I will pose to the class is, how might this new capability be used to promote human understanding and empowerment when using LLMs? SAEs (sparse autoencoders) are too expensive and hard to train to be a complete solution to XAI problems, but they can add depth to a multi-pronged XAI strategy.

    Safety implications: Anthropic’s work on safety is also worth a mention. They identified the ‘syncophantic praise’ feature as part of their work on safety, specifically relevant to this question: could a very powerful AI hide its intentions from humans, possibly by flattering users into complacency? This general direction is especially salient in light of this recent work: Frontier Models are Capable of In-context Scheming.

    Evidence of AI ‘Understanding’? Did interpretability kill the ‘stochastic parrot’? I have been convinced for a while that LLMs must have some internal representations of complex and inter-related concepts. They could not do what they do as one-deep stimulus-response or word-association engines, (‘stochastic parrots’) no matter how many patterns were memorized. The use of complex abstractions, such as those identified by Anthropic, fits my definition of ‘understanding’, although some reserve that term only for human understanding. Perhaps we should just add a qualifier for ‘AI understanding’. This is not a topic that I explicitly cover in my ethics class, but it does come up in discussion of related topics.

    SAE visualization needed. I am still looking for a good visual illustration of how complex features across a deep network are mapped onto to a very thin, very wide SAEs with sparsely represented features. What I have now is the Powerpoint approximation I created for class use, below. Props to Brendan Boycroft for his LLM visualizer, which has helped me understand more about the mechanics of LLMs. https://bbycroft.net/llm

    Author’s depiction of SAE mapping

    Topic 2: Human Centered AI (HCAI)

    What’s new?

    In 2024 it was increasingly apparent that AI will affect every human endeavor and seems to be doing so at a much faster rate than previous technologies such as steam power or computers. The speed of change matters almost more than the nature of change because human culture, values, and ethics do not usually change quickly. Maladaptive patterns and precedents set now will be increasingly difficult to change later.

    What’s important?

    Human-Centered AI needs to become more than an academic interest, it needs to become a well-understood and widely practiced set of values, practices and design principles. Some people and organizations that I like, along with the Anthropic explainability work already mentioned, are Stanford’s Human-Centered AI, Google’s People + AI effort, and Ben Schneiderman’s early leadership and community organizing.

    For my class of working AI engineers, I am trying to focus on practical and specific design principles. We need to counter the dysfunctional design principles I seem to see everywhere: ‘automate everything as fast as possible’, and ‘hide everything from the users so they can’t mess it up’. I am looking for cases and examples that challenge people to step up and use AI in ways that empower humans to be smarter, wiser and better than ever before.

    I wrote fictional cases for class modules on the Future of Work, HCAI and Lethal Autonomous Weapons. Case 1 is about a customer-facing LLM system that tried to do too much too fast and cut the expert humans out of the loop. Case 2 is about a high school teacher who figured out most of her students were cheating on a camp application essay with an LLM and wants to use GenAI in a better way.

    The cases are on separate Medium pages here and here, and I love feedback! Thanks to Sara Bos and Andrew Taylor for comments already received.

    The second case might be controversial; some people argue that it is OK for students to learn to write with AI before learning to write without it. I disagree, but that debate will no doubt continue.

    I prefer real-world design cases when possible, but good HCAI cases have been hard to find. My colleague John (Ian) McCulloh recently gave me some great ideas from examples he uses in his class lectures, including the Organ Donation case, an Accenture project that helped doctors and patients make time-sensitive kidney transplant decision quickly and well. Ian teaches in the same program that I do. I hope to work with Ian to turn this into an interactive case for next year.

    Topic 3: AI governance

    Most people agree that AI development needs to be governed, through laws or by other means, but there’s a lot of disagreement about how.

    What’s new?

    The EU’s AI Act came into effect, giving a tiered system for AI risk, and prohibiting a list of highest-risk applications including social scoring systems and remote biometric identification. The AI Act joins the EU’s Digital Markets Act and the General Data Protection Regulation, to form the world’s broadest and most comprehensive set of AI-related legislation.

    California passed a set of AI governance related laws, which may have national implications, in the same way that California laws on things like the environment have often set precedent. I like this (incomplete) review from the White & Case law firm.

    For international comparisons on privacy, I like DLA Piper‘s website Data Protection Laws of the World.

    What’s Important?

    My class will focus on two things:

    1. How we should evaluate new legislation
    2. How legislation fits into the larger context of AI governance

    How do you evaluate new legislation?

    Given the pace of change, the most useful thing I thought I could give my class is a set of heuristics for evaluating new governance structures.

    Pay attention to the definitions. Each of the new legal acts faced problems with defining exactly what would be covered; some definitions are probably too narrow (easily bypassed with small changes to the approach), some too broad (inviting abuse) and some may be dated quickly.

    California had to solve some difficult definitional problems in order to try to regulate things like ‘Addictive Media’ (see SB-976), ‘AI Generated Media’ (see AB-1836), and to write separate legislation for ‘Generative AI’, (see SB-896). Each of these has some potentially problematic aspects, worthy of class discussion. As one example, The Digital Replicas Act defines AI-generated media as “an engineered or machine-based system that varies in its level of autonomy and that can, for explicit or implicit objectives, infer from the input it receives how to generate outputs that can influence physical or virtual environments.” There’s a lot of room for interpretation here.

    Who is covered and what are the penalties? Are the penalties financial or criminal? Are there exceptions for law enforcement or government use? How does it apply across international lines? Does it have a tiered system based on an organization’s size? On the last point, technology regulation often tries to protect startups and small companies with thresholds or tiers for compliance. But California’s governor vetoed SB 1047 on AI safety for exempting small companies, arguing that “Smaller, specialized models may emerge as equally or even more dangerous”. Was this a wise move, or was he just protecting California’s tech giants?

    Is it enforceable, flexible, and ‘future-proof’? Technology legislation is very difficult to get right because technology is a fast-moving target. If it is too specific it risks quickly becoming obsolete, or worse, hindering innovations. But the more general or vague it is, the less enforceable it may be, or more easily ‘gamed’. One strategy is to require companies to define their own risks and solutions, which provides flexibility, but will only work if the legislature, the courts and the public later pay attention to what companies actually do. This is a gamble on a well-functioning judiciary and an engaged, empowered citizenry… but democracy always is.

    How does legislation fit into the bigger picture of AI governance?

    Not every problem can or should be solved with legislation. AI governance is a multi-tiered system. It includes the proliferation of AI frameworks and independent AI guidance documents that go further than legislation should, and provide non-binding, sometimes idealistic goals. A few that I think are important:

    Other miscellaneous topics

    Here’s some other news items and topics I am integrating into my class, some of which are new to 2024 and some are not. I will:

    Thanks for reading! I always appreciate making contact with other people teaching similar courses or with deep knowledge of related areas. And I also always appreciate Claps and Comments!


    What I’m Updating in My AI Ethics Class for 2025 was originally published in Towards Data Science on Medium, where people are continuing the conversation by highlighting and responding to this story.

    Originally appeared here:
    What I’m Updating in My AI Ethics Class for 2025

    Go Here to Read this Fast! What I’m Updating in My AI Ethics Class for 2025

  • Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

    Take a look at what Apple is going to bring us in 2025 — and while you’re doing that, take a look at everything Apple TV+ has for you right now as it launches its first-ever free weekend.

    Store interior with people milling about and an Apple TV+ promotional sign displaying various TV show posters.
    An original Apple TV+ in-store promotion from 2010 — image credit: William Gallagher

    Apple TV+ has never gone completely free before. There have been so many different free trials when you bought various devices, but this is the very first time that anyone can watch anything on Apple TV+, as long as they watch from Friday to Sunday, January 3 to 5, 2025.

    Continue Reading on AppleInsider | Discuss on our Forums

    Go Here to Read this Fast!

    Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

    Originally appeared here:

    Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

  • Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

    Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

    Take a look at what Apple is going to bring us in 2025 — and while you’re doing that, take a look at everything Apple TV+ has for you right now as it launches its first-ever free weekend.

    Store interior with people milling about and an Apple TV+ promotional sign displaying various TV show posters.
    An original Apple TV+ in-store promotion from 2010 — image credit: William Gallagher

    Apple TV+ has never gone completely free before. There have been so many different free trials when you bought various devices, but this is the very first time that anyone can watch anything on Apple TV+, as long as they watch from Friday to Sunday, January 3 to 5, 2025.

    Continue Reading on AppleInsider | Discuss on our Forums

    Go Here to Read this Fast!

    Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

    Originally appeared here:

    Apple’s 2025 releases and free Apple TV+ for the weekend, on the AppleInsider Podcast

  • Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Snag today’s lowest iPad 10th Generation price at Amazon, with a $70 discount in effect heading into the weekend.

    iPad 10th Generation on desk with plants, headphones, orange HomePod mini, and a lamp in the background; 'Deal of the Week' sign upper left.
    Save $70 with this iPad 10th Gen markdown

    Take advantage of a fresh price drop on Apple’s iPad 10th Generation, with the standard 64GB Wi-Fi model falling to $279 at Amazon and Best Buy.

    Get the $279 deal

    Continue Reading on AppleInsider | Discuss on our Forums

    Go Here to Read this Fast! Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Originally appeared here:
    Grab the iPad 10th Gen for just $279 at Amazon this weekend

  • Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Snag today’s lowest iPad 10th Generation price at Amazon, with a $70 discount in effect heading into the weekend.

    iPad 10th Generation on desk with plants, headphones, orange HomePod mini, and a lamp in the background; 'Deal of the Week' sign upper left.
    Save $70 with this iPad 10th Gen markdown

    Take advantage of a fresh price drop on Apple’s iPad 10th Generation, with the standard 64GB Wi-Fi model falling to $279 at Amazon and Best Buy.

    Get the $279 deal

    Continue Reading on AppleInsider | Discuss on our Forums

    Go Here to Read this Fast! Grab the iPad 10th Gen for just $279 at Amazon this weekend

    Originally appeared here:
    Grab the iPad 10th Gen for just $279 at Amazon this weekend

  • Apple Watch health tips to kick off the new year

    Apple Watch health tips to kick off the new year

    Going into the new year, here are our best tips to maximize your health with the help of any Apple Watch.

    Two hands hold smartwatches with metal bands, displaying digital time and fitness stats, against a light background with a small potted succulent nearby.
    Apple Watch is a great health companion

    Like its predecessors, the latest Apple Watch Series 10 is an incredible fitness tool that is packed with sensors that give you immediate health and fitness information. Whether your goal is to lose weight, live healthier, or something else — any Apple Watch can help.

    That’s always true, but right now at the start of the year, many users are opening new Apple Watches from the holidays. Plus many of those, and existing users, will be starting off their new year resolutions which always include getting fitter.

    Continue Reading on AppleInsider | Discuss on our Forums

    Go Here to Read this Fast! Apple Watch health tips to kick off the new year

    Originally appeared here:
    Apple Watch health tips to kick off the new year