by time.com — BY NIK POPLI — Ammaar Reshi was playing around with ChatGPT, an AI-powered chatbot from OpenAI when he started thinking about the ways artificial intelligence could be used to make a simple children’s book to give to his friends. Just a couple of days later, he published a 12-page picture book, printed it, and started selling it on Amazon without ever picking up a pen and paper.
Reshi, a product design manager from the San Francisco Bay Area, gathered illustrations from Midjourney, a text-to-image AI tool that launched this summer, and took story elements from a conversation he had with the AI-powered ChatGPT about a young girl named Alice. “Anyone can use these tools,” Reshi tells TIME. “It’s easily and readily accessible, and it’s not hard to use either.” The feat, which Reshi publicized in a viral Twitter thread, is a testament to the incredible advances in AI-powered tools like ChatGPT—which took the internet by storm two weeks ago with its uncanny ability to mimic human thought and writing. But the book, Alice and Sparkle, also renewed a fierce debate about the ethics of AI-generated art. Many argued that the technology preys on artists and other creatives—using their hard work as source material, while raising the specter of replacing them.
His experiment creating an AI-generated book in just one weekend shows that artificial intelligence might be able to accomplish tasks faster and more efficiently than any human person can—sort of. The book was far from perfect. The AI-generated illustrations had a number of issues: some fingers looked like claws, objects were floating, and the shadowing was off in some areas. Normally, illustrations in children’s books go through several rounds of revisions—but that’s not always possible with AI-generated artwork on Midjourney, where users type a series of words and the bot spits back an image seconds later. Alice and Sparkle follows a young girl who builds her own artificial intelligence robot that becomes self aware and capable of making its own decisions. Reshi has sold about 70 copies through Amazon since Dec. 4, earning royalties of less than $200. He plans to donate additional copies to his local library. Reshi’s quixotic project drew praise from many users for its ingenuity. But many artists also strongly criticized both his process and the product. To his critics, the speed and ease with which Reshi created Alice and Sparkle exemplifies the ethical concerns of AI-generated art. Artificial intelligence systems like Midjourney are trained using datasets of millions of images that exist across the Internet, then teaching algorithms to recognize patterns in those images and generate new ones. That means any artist who uploads their work online could be feeding the algorithm without their consent. Many claim this amounts to a high-tech form of plagiarism that could seriously harm human artists in the near future. Reshi’s original tweet promoting his book received more than 6 million impressions and 1,300 replies, many of which came from book illustrators claiming artists should be paid or credited if their work is used by AI.
by venturebeat.com — Ben Dickson — For decades, we have personified our devices and applications with verbs such as “thinks,” “knows” and “believes.” And in most cases, such anthropomorphic descriptions are harmless. But we’re entering an era in which we must be careful about how we talk about software, artificial intelligence (AI) and, especially, large language models (LLMs), which have become impressively advanced at mimicking human behavior while being fundamentally different from the human mind. It is a serious mistake to unreflectively apply to artificial intelligence systems the same intuitions that we deploy in our dealings with each other, warns Murray Shanahan, professor of Cognitive Robotics at Imperial College London and a research scientist at DeepMind, in a new paper titled, “Talking About Large Language Models.” And to make the best use of the remarkable capabilities AI systems possess, we must be conscious of how they work and avoid imputing to them capacities they lack.
Humans vs. LLMs
“It’s astonishing how human-like LLM-based systems can be, and they are getting better fast. After interacting with them for a while, it’s all too easy to start thinking of them as entities with minds like our own,” Shanahan told VentureBeat. “But they are really rather an alien form of intelligence, and we don’t fully understand them yet. So we need to be circumspect when incorporating them into human affairs.” Human language use is an aspect of collective behavior. We acquire language through our interactions with our community and the world we share with them. “As an infant, your parents and carers offered a running commentary in natural language while pointing at things, putting things in your hands or taking them away, moving things within your field of view, playing with things together, and so on,” Shanahan said. “LLMs are trained in a very different way, without ever inhabiting our world.” LLMs are mathematical models that represent the statistical distribution of tokens in a corpus of human-generated text (tokens can be words, parts of words, characters or punctuations). They generate text in response to a prompt or question, but not in the same way that a human would do. Shanahan simplifies the interaction with an LLM as such: “Here’s a fragment of text. Tell me how this fragment might go on. According to your model of the statistics of human language, what words are likely to come next?”