Posts with «author_name|andrew tarantola» label

The US Senate and Silicon Valley reconvene for a second AI Insight Forum

Senator Charles Schumer (D-NY) once again played host to Silicon Valley’s AI leaders on Tuesday as the US Senate reconvened its AI Insights Forum for a second time. On the guest list this go around: manifesto enthusiast Marc Andreessen and venture capitalist John Doerr, as well as Max Tegmark of the Future of Life Institute and NAACP CEO Derrick Johnson. On the agenda: “the transformational innovation that pushes the boundaries of medicine, energy, and science, and the sustainable innovation necessary to drive advancements in security, accountability, and transparency in AI,” according to a release from Sen. Schumer’s office.

Upon exiting the meeting Tuesday, Schumer told the assembled press, "it is clear that American leadership on AI can’t be done on the cheap. Almost all of the experts in today’s Forum called for robust, sustained federal investment in private and public sectors to achieve our goals of American-led transformative and sustainable innovation in AI. 

Per National Security AI Commission estimates, paying for that could cost around $32 billion a year. However, Schumer believes that those funding challenges can be addressed by "leveraging the private sector by employing new and innovative funding mechanisms – like the Grand Challenges prize idea." 

"We must prioritize transformational innovation, to help create new vistas, unlock new cures, improve education, reinforce national security, protect the global food supply, and more," Schumer remarked. But in doing so, we must act sustainably in order to minimize harms to workers, civil society and the environment. "We need to strike a balance between transformational and sustainable innovation," Schumer said. "Finding this balance will be key to our success."

Senators Brian Schatz (D-HI) and John Kennedy (R-LA) also got in on the proposed regulatory action Tuesday, introducing legislation that would provide more transparency on AI-generated content by requiring clear labeling and disclosures. Such technology could resemble the Content Credentials tag that the C2PA and CAI industry advocacy groups are developing.

"Our bill is simple," Senator Schatz said in a press statement. "If any content is made by artificial intelligence, it should be labeled so that people are aware and aren’t fooled or scammed.”

The Schatz-Kennedy AI Labeling Act, as they're calling it, would require generative AI system developers to clearly and conspicuously disclose AI-generated content to users. Those developers, and their licensees, would also have to take "reasonable steps" to prevent "systematic publication of content without disclosures." The bill would also establish a working group to create non-binding technical standards to help social media platforms automatically identify such content as well.

“​​It puts the onus where it belongs: on the companies and not the consumers,” Schatz said on the Senate floor Tuesday. “Labels will help people to be informed. They will also help companies using AI to build trust in their content.”

Tuesday’s meeting follows the recent introduction of new AI legislation, dubbed the Artificial Intelligence Advancement Act of 2023 (S. 3050). Senators Martin Heinrich (D-NM), Mike Rounds (R-SD), Charles Schumer (D-NY) and Todd Young (R-IN) all co-sponsored the bill. The bill proposes AI bug bounty programs and would require a vulnerability analysis study for AI-enabled military applications. It’s passage into law would also launch a report into AI regulation in the financial services industry (which the head of the SEC had recently been lamenting) as well as a second report on data sharing and coordination.

“It’s frankly a hard challenge,” SEC Chairman Gary Gensler told The Financial Times recently, speaking on the challenges the financial industry faces in AI adoption and regulation. “It’s a hard financial stability issue to address because most of our regulation is about individual institutions, individual banks, individual money market funds, individual brokers; it’s just in the nature of what we do.”

"Working people are fighting back against artificial intelligence and other technology used to eliminate workers or undermine and exploit us," AFL-CIO President Liz Shuler said at the conclusion of Tuesday's forum. "If we fail to involve workers and unions across the entire innovation process, AI will curtail our rights, threaten good jobs and undermine our democracy. But the responsible adoption of AI, properly regulated, has the potential to create opportunity, improve working conditions and build prosperity."

The forums are part of Senator Schumer’s SAFE Innovation Framework, which his office debuted in June. “The US must lead in innovation and write the rules of the road on AI and not let adversaries like the Chinese Communist Party craft the standards for a technology set to become as transformative as electricity,” the program announcement reads.

While Andreesen calls for AI advancement at any cost and Tegmark continues to advocate for a developmental “time out,” rank and file AI industry workers are also fighting to make their voices heard ahead of the forum. On Monday, a group of employees from two dozen leading AI firms published an open letter to Senator Schumer, demanding Congress take action to safeguard their livelihoods from the “dystopian future” that Andreessen’s screed, for example, would require.

“Establishing robust protections related to workplace technology and rebalancing power between workers and employers could reorient the economy and tech innovation toward more equitable and sustainable outcomes,” the letter authors argue.

Senator Ed Markey (D-MA) and Representative Pramila Jayapal (WA-07) had, the previous month, called on leading AI companies to “answer for the working conditions of their data workers, laborers who are often paid low wages and provided no benefits but keep AI products online.”

"We covered a lot of good ground today, and I think we’ll all be walking out of the room with a deeper understanding of how to approach American-led AI innovation," Schumer said Tueseay. "We’ll continue this conversation in weeks and months to come – in more forums like this and committee hearings in Congress – as we work to develop comprehensive, bipartisan AI legislation."

This article originally appeared on Engadget at https://www.engadget.com/the-us-senate-and-silicon-valley-reconvene-for-a-second-ai-insight-forum-143128622.html?src=rss

Adult film star Riley Reid launches Clona.AI, a sexting chatbot platform

Adult film icon and media investor Riley Reid aims to bring the transformational capabilities of generative AI to adult entertainment with an online platform where users can chat with digital versions of content creators. But unlike other, scuzzier adult chatbots, Clona.AI’s avatars are trained with explicit consent of the models’ creators who have direct input in what the “AI companions” will, and won’t, talk about.

For $30 a month, fans and subscribers will be able to hold “intimate conversations” with digital versions of their favorite adult stars, content creators and influencers. The site’s roster currently includes Reid herself and Lena the Plug. A free tier is also available but offers just five chat messages per month. 

“The reality is, AI is coming, and if it's not Clona, it’s somebody else,” Reid told 404 Media. “When [other people] use deepfakes or whatever — if I'm not partnering up with it, then someone else is going to steal my likeness and do it without me. So being presented with this opportunity, I was so excited because I felt like I had a chance to be a part of society's technological advances.”

Clona uses Meta’s Llama 2 large language model as a base, then heavily refines and retrains it to reflect the personality of the person it’s based on. Reid explains that her model was first trained on a variety of her online media including interviews, podcast appearances and YouTube videos (in addition to some of her x-rated work) before further fine tuning its response by having the AI chat with Reid herself.

“I’ll be able to see how it responds to users, and edit it to be like ‘no, I would have said it more like this,’’’ Reid said. “But in the beginning my focus was on things like making sure it had my dogs’ names right, making sure I was fact-checking it.”

While the AI companion will be capable talking dirty, how dirty that gets depends on the actor’s preferences, not the user’s. Reid notes that her model, for example, will not discuss physically dangerous sex acts with users. "I don't know if the tech team thought about the sounding guys, but I was like, I thought about them,” she said.

Generative AI technology has shown tremendous potential in creating digital clones of deceased celebrities and recording artists. The process requires little more than the celeb’s permission (or that of their estate) and a sufficiently large corpus of their vocal or video recordings. However, we’ve already also seen that technology be misused in deepfake pornography and shady dental advertising. Unscrupulous data scraping practices on the public web (data which is then used to train LLMs) has also raised difficult questions regarding modern copyright laws, copyright infringement and Grammy award eligibility.

Still, Reid remains optimistic about the historically proven resilience of the sex industry. “I feel like we're gonna be a huge part of AI adapting into our society, because porn is always like that,” Reid said. “It’s what it did with the internet. And the porn world has seen so many advances in technology.”

This article originally appeared on Engadget at https://www.engadget.com/adult-film-star-riley-reid-launches-clonaai-a-sexting-chatbot-platform-000509221.html?src=rss

Qualcomm brings on-device AI to mobile and PC

Qualcomm is no stranger in running artificial intelligence and machine learning systems on-device and without an internet connection. They’ve been doing it with their camera chipsets for years. But on Tuesday at Snapdragon Summit 2023, the company announced that on-device AI is finally coming to mobile devices and Windows 11 PCs as part of the new Snapdragon 8 Gen 3 and X Elite chips.

Both chipsets were built from the ground up with generative AI capabilities in mind and are able to support a variety of large language models (LLM), language vision models (LVM), and transformer network-based automatic speech recognition (ASR) models, up to 10 billion parameters for the SD8 gen 3 and 13 billion parameters for the X Elite, entirely on-device. That means you’ll be able to run anything from Baidu’s ERNIE 3.5 to OpenAI’s Whisper, Meta's Llama 2 or Google’s Gecko on your phone or laptop, without an internet connection. Qualcomm’s chips are optimized for voice, text and image inputs.

“It's important to have a wide array of support underneath the hood for these models to be running and therefore heterogeneous compute is extremely important,” Durga Malladi, SVP & General Manager, Technology Planning & Edge Solutions at Qualcomm, told reporters at a prebriefing last week. “We have state-of-the-art CPU, GPU, and NPU (Neural Processing Unit) processors that are used concurrently, as multiple models are running at any given point in time.”

The Qualcomm AI Engine is comprised of the Oryon CPU, the Adreno GPU and Hexagon NPU. Combined, they handle up to 45 TOPS (trillions of operations per second) and can crunch 30 tokens per second on laptops, 20 tokens per second on mobile devices — tokens being the basic text/data unit that LLMs can process/generate off of. The chipsets use Samsung’s 4.8GHz LP-DDR5x DRAM for their memory allocation.

Qualcomm

“Generative AI has demonstrated the ability to take very complex tasks, solve them and resolve them in a very efficient manner,” he continued. Potential use cases could include meeting and document summarization or email drafting for consumers, and prompt-based computer code or music generation for enterprise applications, Malladi noted.

Or you could just use it to take pretty pictures. Qualcomm is integrating its previous work with edge AI, Cognitive ISP. Devices using these chipsets will be able to edit photos in real-time and in as many as 12 layers. They'll also be able to capture clearer images in low light, remove unwanted objects from photos (a la Google’s Magic Eraser) or expand image backgrounds. User scan even watermark their shots as being real and not AI generated, using Truepic photo capture.

Having an AI that lives primarily on your phone or mobile device, rather than in the cloud, will offer users myriad benefits over the current system. Much like enterprise AIs that take a general model (e.g. GPT-4) and tune it using a company’s internal data to provide more accurate and on-topic answers, a locally-stored AI “over time… gradually get personalized,” Malladi said, “in the sense that… the assistant gets smarter and better, running on the device in itself.”

What’s more, the inherent delay present when the model has to query the cloud for processing or information doesn’t exist when all of the assets are local. As such, both the X Elite and SD8 gen 3 are capable of not only running Stable Diffusion on-device but generating images in less than 0.6 seconds.

The capacity to run bigger and more capable models, and interact with those models using our speaking words instead of our typing words, could ultimately prove the biggest boon to consumers. “There's a very unique way in which we start interfacing the devices and voice becomes a far more natural interface towards these devices — as well in addition to everything else,” Malladi said. “We believe that it has the potential to be a transformative moment, where we start interacting with devices in a very different way compared to what we've done before.”

Mobile devices and PCs are just the start for Qualcomm’s on-device AI plans. The 10-13 billion parameter limit is already moving towards 20 billion-plus parameters as the company develops new chip iterations. “These are very sophisticated models,” Malladi commented. “The use cases that you build on this are quite impressive.”

“When you start thinking about ADAS (Advanced Driver Assistance Systems) and you have multi-modality [data] coming in from multiple cameras, IR sensors, radar, lidar — in addition to voice, which is the human that is inside the vehicle in itself,” he continued. “The size of that model is pretty large, we're talking about 30 to 60 billion parameters already.” Eventually, these on-device models could approach 100 billion parameters or more, according to Qualcomm’s estimates.

This article originally appeared on Engadget at https://www.engadget.com/qualcomm-brings-on-device-ai-to-mobile-and-pc-190030938.html?src=rss

NASA's John Mather keeps redefining our understanding of the cosmos

Space isn't hard only on account of the rocket science. The task of taking a NASA mission from development and funding through construction and launch — all before we even use the thing for science — can span decades. Entire careers have been spent putting a single satellite into space. Nobel-winning NASA physicist John Mather, mind you, has already helped send up two.

In their new book, Inside the Star Factory: The Creation of the James Webb Space Telescope, NASA's Largest and Most Powerful Space Observatory, author Christopher Wanjek and photographer Chris Gunn take readers on a behind the scenes tour of the James Webb Space Telescope's own journey from inception to orbit. Weaving examinations of the radical imaging technology that enables us to peer deeper into the early universe than ever before with profiles of the researchers, advisors, managers, engineers and technicians that made it possible through three decades of effort. In this week's Hitting the Books excerpt, a look at JWST project scientist John Mather and his own improbable journey from rural New Jersey to NASA. 

MIT Press

Excerpted from “Inside the Star Factory: The Creation of the James Webb Space Telescope, NASA's Largest and Most Powerful Space Observatory” Copyright © 2023 by Chris Gunn and Christopher Wanjek. Used with permission of the publisher, MIT Press.


John Mather, Project Scientist 

— The steady hand in control 

John Mather is a patient man. His 2006 Nobel Prize in Physics was thirty years in the making. That award, for unswerving evidence of the Big Bang, was based on a bus-sized machine called COBE — yet another NASA mission that almost didn’t happen. Design drama? Been there. Navigate unforeseen delays? Done that. For NASA to choose Mather as JWST Project Scientist was pure prescience. 

Like Webb, COBE — the Cosmic Background Explorer — was to be a time machine to reveal a snapshot of the early universe. The target era was just 370,000 years after the Big Bang, when the universe was still a fog of elementary particles with no discernable structure. This is called the epoch of recombination, when the hot universe cooled to a point to allow protons to bind with electrons to form the very first atoms, mostly hydrogen with a sprinkling of helium and lithium. As the atoms formed, the fog lifted, and the universe became clear. Light broke through. That ancient light, from the Big Bang itself, is with us today as remnant microwave radiation called the cosmic microwave background. 

Tall but never imposing, demanding but never mean, Mather is a study in contrasts. His childhood was spent just a mile from the Appalachian Trail in rural Sussex County, New Jersey, where his friends were consumed by earthly matters such as farm chores. Yet Mather, whose father was a specialist in animal husbandry and statistics, was more intrigued by science and math. At age six he grasped the concept of infinity when he filled up a page in his notebook with a very large number and realized he could go on forever. He loaded himself up with books from a mobile library that visited the farms every couple of weeks. His dad worked for Rutgers University Agriculture Experiment Station and had a laboratory on the farm with radioisotope equipment for studying metabolism and liquid nitrogen tanks with frozen bull semen. His dad also was one of the earliest users of computers in the area, circa 1960, maintaining milk production records of 10,000 cows on punched IBM cards. His mother, an elementary school teacher, was quite learned, as well, and fostered young John’s interest in science.

A chance for some warm, year-round weather ultimately brought Mather in 1968 to University of California, Berkeley, for graduate studies in physics. He would fall in with a crowd intrigued by the newly detected cosmic microwave background, discovered by accident in 1965 by radio astronomers Arno Penzias and Robert Wilson. His thesis advisor devised a balloon experiment to measure the spectrum, or color, of this radiation to see if it really came from the Big Bang. (It does.) The next obvious thing was to make a map of this light to see, as theory suggested, whether the temperature varied ever so slightly across the sky. And years later, that’s just what he and his COBE team found: anisotropy, an unequal distribution of energy. These micro-degree temperature fluctuations imply matter density fluctuations, sufficient to stop the expansion, at least locally. Through the influence of gravity, matter would pool into cosmic lakes to form stars and galaxies hundreds of millions of years later. In essence, Mather and his team captured a sonogram of the infant universe. 

Yet the COBE mission, like Webb, was plagued with setbacks. Mather and the team proposed the mission concept (for a second time) in 1976. NASA accepted the proposal but, that year, declared that this satellite and most others from then on would be delivered to orbit by the Space Shuttle, which itself was still in development. History would reveal the foolishness of such a plan. Mather understood immediately. This wedded the design of COBE to the cargo bay of the unbuilt Shuttle. Engineers would need to meet precise mass and volume requirements of a vessel not yet flown. More troublesome, COBE required a polar orbit, difficult for the Space Shuttle to deliver. The COBE team was next saddled with budget cuts and compromises in COBE’s design as a result of cost overruns of another pioneering space science mission, the Infrared Astronomical Satellite, or IRAS. Still, the tedious work continued of designing instruments sensitive enough to detect variations of temperatures just a few degrees above absolute zero, about −270°C. From 1980 onward, Mather was consumed by the creation of COBE all day every day. The team needed to cut corners and make risky decisions to stay within budget. News came that COBE was to be launched on the Space Shuttle mission STS-82-B in 1988 from Vandenberg Air Force Base. All systems go.

Then the Space Shuttle Challenger exploded in 1986, killing all seven of its crew. NASA grounded Shuttle flights indefinitely. COBE, now locked to Shuttle specifications, couldn’t launch on just any other rocket system. COBE was too large for a Delta rocket at this point; ironically, Mather had the Delta in mind in his first sketch in 1974. The team looked to Europe for a launch vehicle, but this was hardly an option for NASA. Instead, the project managers led a redesign to shave off hundreds of pounds, to slim down to a 5,000-pound launch mass, with fuel, which would just make it within the limits of a Delta by a few pounds. Oh, and McDonnell Douglas had to build a Delta rocket from spare parts, having been forced to discontinue the series in favor of the Space Shuttle. 

The team worked around the clock over the next two years. The final design challenge was ... wait for it ... a sunshield that now needed to be folded into the rocket and spring-released once in orbit, a novel approach. COBE got the greenlight to launch from Vandenberg Air Force Base in California, the originally desired site because it would provide easier access to a polar orbit compared to launching a Shuttle from Florida. Launch was set for November 1989. COBE was delivered several months before. 

Then, on October 17, the California ground shook hard. A 6.9-magnitude earthquake struck Santa Cruz County, causing widespread damage to structures. Vandenberg, some 200 miles south, felt the jolt. As pure luck would have it, COBE was securely fastened only because two of the engineers minding it secured it that day before going off to get married. The instrument suffered no damage and launched successfully on November 18. More drama came with the high winds on launch day. Myriad worries followed in the first weeks of operation: the cryostat cooled too quickly; sunlight reflecting off of Antarctic ice played havoc with the power system; trapped electrons and protons in the Van Allen belts disrupted the functioning of the electronics; and so on. 

All the delays, all the drama, faded into a distant memory for Mather as the results of the COBE experiment came in. Data would take four years to compile. But the results were mind-blowing. The first result came weeks after launch, when Mather showed the spectrum to the American Astronomical Society and received a standing ovation. The Big Bang was safe as a theory. Two years later, at an April 1992 meeting of the American Physical Society, the team showed their first map. Data matched theory perfectly. This was the afterglow of the Big Bang revealing the seeds that would grow into stars and galaxies. Physicist Stephen Hawking called it “the most important discovery of the century, if not of all time.” 

Mather spoke humbly of the discovery at his Nobel acceptance speech in 2006, fully crediting his remarkable team and his colleague George Smoot, who shared the prize with him that year. But he didn’t downplay the achievement. He noted that he was thrilled with the now broader “recognition that our work was as important as people in the professional astronomy world have known for so long.” 

Mather maintains that realism today. While concerned about delays, threats of cancellation, cost overruns, and not-too-subtle animosity in the broader science community over the “telescope that ate astronomy,” he didn’t let this consume him or his team. “There’s no point in trying to manage other people’s feelings,” he said. “Quite a lot of the community opinion is, ‘well, if it were my nickel, I’d spend it differently.’ But it isn’t their nickel; and the reason why we have the nickel in the first place is because NASA takes on incredibly great challenges. Congress approved of us taking on great challenges. And great challenges aren’t free. My feeling is that the only reason why we have an astronomy program at NASA for anyone to enjoy — or complain about — is that we do astonishingly difficult projects. We are pushing to the edge of what is possible.” 

Webb isn’t just a little better than the Hubble Space Telescope, Mather added; it’s a hundred times more powerful. Yet his biggest worry through mission design was not the advanced astronomy instruments but rather the massive sunshield, which needed to unfold. All instruments and all the deployment mechanisms had redundancy engineered into them; there are two or more ways to make them work if the primary method fails. But that’s not the only issue with a sunshield. It would either work or not work. 

Now Mather can focus completely on the science to be had. He expects surprises; he’d be surprised if there were no surprises. “Just about everything in astronomy comes as a surprise,” he said. “When you have new equipment, you will get a surprise.” His hunch is that Webb might reveal something weird about the early universe, perhaps an abundance of short-lived objects never before seen that say something about dark energy, the mysterious force that seems to be accelerating the expansion of the universe, or the equally mysterious dark matter. He also can’t wait until Webb turns its cameras to Alpha Centauri, the closest star system to Earth. What if there’s a planet there suitable for life? Webb should have the sensitivity to detect molecules in its atmosphere, if present. 

“That would be cool,” Mather said. Hints of life from the closest star system? Yes, cool, indeed.

This article originally appeared on Engadget at https://www.engadget.com/inside-the-star-factory-chris-gunn-christopher-wanjek-mit-press-143046496.html?src=rss

Tesla begins Cybertruck deliveries on November 30

After slogging through years of delays and redesigns, the Tesla Cybertruck can finally be seen on public roads this holiday season, the company announced. Deliveries of the long-awaited luxury EV SUV will begin to select customers starting November 30, before the vehicle enters full production next year at its Texas Gigafactory.  

Q3 Shareholder Update → https://t.co/sXBSeLibSL

Highlights

Cybertruck production remains on track for later this year, with first deliveries scheduled for November 30th at Giga Texas.

Production of our higher density 4680 cell is progressing as planned & we continue building… pic.twitter.com/FqpseLujaA

— Tesla (@Tesla) October 18, 2023

For its existing model lines, Tesla's production and deliveries are both down this quarter, about seven percent or roughly 30,000 units compared to Q2, but still significantly higher year over year, up ~100,000 units, over 2022. The EV automaker has slashed the prices on its vehicles repeatedly this year, first in March, then again in September (taking a full 20 percent off the MSRP at the time) and once more in early October

The Model X, for example, began 2023 retailing for $120,990 — it currently lists for $79,990. The models S (now $74,990), Y ($52,490, down 24 percent from January) and 3 ($38,990, down 17 percent) have all seen similar price drops. In all, Tesla reports its cost of goods sold per vehicle decreased to ~$37,500 in Q3.

Musk had previously explained his willingness to drop prices and endure reduced margins if it translates to increased sales volume. “I think it does make sense to sacrifice margins in favor of making more vehicles,” he said in July. 

“A sequential decline in volumes was caused by planned downtimes for factory upgrades, as discussed on the most recent earnings call. Our 2023 volume target of around 1.8 million vehicles remains unchanged,” Tesla wrote in an October press statement. The company delivered some 435,059 vehicles globally in Q3. 

The company continues to increase its investments in AI development as well, having "more than doubled" the amount of processing power it dedicates to training its vehicular and Optimus robot AI systems, compared to Q2. The Optimus itself is reportedly receiving hardware upgrades and is being trained via AI, rather than "hard-coded" software. Additionally, the company announced that all US and Canadian Hertz rentals will have access to the Tesla App, allowing them to use their phones as key fobs. 

This is a developing story. Please check back for updates.

This article originally appeared on Engadget at https://www.engadget.com/tesla-begins-cybertruck-deliveries-on-november-30-210430697.html?src=rss

Baidu's CEO says its ERNIE AI 'is not inferior in any aspect to GPT-4'

ERNIE, Baidu’s answer to ChatGPT, has “achieved a full upgrade,” company CEO Robin Li told the assembled crowd at the Baidu World 2023 showcase on Tuesday, “with drastically improved performance in understanding, generation, reasoning, and memory.”

During his keynote address, Li demonstrated improvements to those four core capabilities on-stage by having the AI create a multimodal car commercial in a few minutes based on a short text prompt , solve complex geometry problems and progressively iterate the plot for a short story on the spot. The fourth-gen generative AI system “is not inferior in any aspect to GPT-4,” he continued.

ERNIE 4.0 will offer an “improved” search experience resembling that of Google’s SGE, aggregating and summarizing information pulled from the wider web and distilled into a generated response.The system will be multimodal, providing answers as text, images or animated graphs through an “interactive chat interface for more complex searches, enabling users to iteratively refine their queries until reaching the optimal answer, all in one search interface,” per the company’s press. What’s more, the AI will be able to recommend “highly customized” content streams based on previous interactions with the user.

Similar to ChatGPT Enterprise, ERNIE’s new Generative Business Intelligence will offer a more finely-tuned and secure model trained on each client’s individual data silo. ERNIE 4.0 will also be capable of, “conducting academic research, summarizing key information, creating documents, and generating slideshow presentations” as well as enable users to search and retrieve files using text and voice prompts.

Baidu is following the example set by the rest of the industry and has announced plans to put its generative AI in every app and service it can manage. The company has already integrated some of the AI’s functions into Baidu Maps, including navigation, ride hailing and hotel bookings. It is also offering “ow-threshold access and productivity tools” to help individuals and enterprises develop API plugins for the Baidu Qianfan Foundation Model Platform.

Baidu had already been developing its ERNIE large language model for a number of years prior to the debut of ChatGPT in 2022, though its knowledge-base is focused primarily on the Chinese market. Baidu released ERNIE Bot, it’s answer to ChatGPT, this March with some 550 billion facts packed into its knowledge graph, though it wasn’t until this August that it rolled out to the general public.

Baidu’s partner startups also showed off new product series that will integrate the AI’s functionality during the event, including a domestic robot, an All-in-One learning machine and a smart home speaker.

This article originally appeared on Engadget at https://www.engadget.com/baidus-ceo-says-its-ernie-ai-is-not-inferior-in-any-aspect-to-gpt-4-162333722.html?src=rss

Honda to test its Autonomous Work Vehicle at Toronto's Pearson Airport

While many of the flashy, marquee mobility and transportation demos that go on at CES tend to be of the more... aspirational variety, Honda's electric cargo hauler, the Autonomous Work Vehicle (AWV), could soon find use on airport grounds as the robotic EV trundles towards commercial operations. 

Honda first debuted the AWV as part of its CES 2018 companion mobility demonstration, then partnered with engineering firm Black & Veatch to further develop the platform. The second-generation AWV was capable of being remotely piloted or following a preset path while autonomously avoiding obstacles. It could carry nearly 900 pounds of sutff onboard and atow another 1,600 pounds behind it, both on-road and off-road. Those second-gen prototypes spent countless hours ferrying building materials back and forth across a 1,000-acre solar panel construction worksite, both individually and in teams, as part of the development process. 

This past March, Honda unveiled the third-generation AWV with a higher carrying capacity, higher top speed, bigger battery and better obstacle avoidance. On Tuesday, Honda revealed that it is partnering with the Greater Toronto Airports Authority to test its latest AWV at the city's Pearson Airport. 

The robotic vehicles will begin their residencies by driving the perimeters of airfields, using mounted cameras and an onboard AI, checking fences and reporting any holes or intrusions. The company is also considering testing the AWV as a FOD (foreign object debris) tool to keep runways clear, as an aircraft component hauler, people mover or baggage cart tug. 

The AWV is just a small part of Honda's overall electrification efforts. The automaker is rapidly shifting its focus from internal combustion to e-motors with plans to release a fully-electric mid-size SUV, as well as nearly a dozen EV motorcycle models by 2025, and develop an EV sedan with Sony. Most importantly, however, the Motocompatco is making a comeback

This article originally appeared on Engadget at https://www.engadget.com/honda-to-test-its-autonomous-work-vehicle-at-torontos-pearson-airport-153025911.html?src=rss

Hitting the Books: Voice-controlled AI copilots could lead to safer flights

Siri and Alexa were only the beginning. As voice recognition and speech synthesis technologies continue to mature, the days of typing on keyboards to interact with the digital world around us could be coming to an end — and sooner than many of us anticipated. Where today's virtual assistants exist on our mobile devices and desktops to provide scripted answers to specific questions, the LLM-powered generative AI copilots of tomorrow will be there, and everywhere else too. This is the "voice-first" future Tobias Dengel envisions in his new book, The Sound of the Future: The Coming Age of Voice Technology.

Using a wide-ranging set of examples, and applications in everything from marketing, sales and customer service to manufacturing and logistics, Dengel walks the reader through how voice technologies can revolutionize the ways in which we interact with the digital world. In the excerpt below, Dengel discusses voice technology might expand its role in the aviation industry, even after the disastrous outcome of its early use in the Boeing 737 MAX.       

PublicAffairs

Excerpted from THE SOUND OF THE FUTURE: The Coming Age of Voice Technology by Tobias Dengel with Karl Weber. Copyright © 2023. Available from PublicAffairs, an imprint of Hachette Book Group, Inc.


REDUCING THE BIGGEST RISKS: MAKING FLYING SAFER

Some workplaces involve greater risks than others. Today’s technology-driven society sometimes multiplies the risks we face by giving ordinary people control over once-incredible amounts of power, in forms that range from tractor trailers to jet airplanes. People carrying out professional occupations that involve significant risks on a daily basis will also benefit from the safety edge that voice provides — as will the society that depends on these well-trained, highly skilled yet imperfect human beings.

When the Boeing 737 MAX airliner was rolled out in 2015, it featured a number of innovations, including distinctive split-tip winglets and airframe modifications that affected the jumbo jet’s aerodynamic characteristics. A critical launch goal for Boeing was to enable commercial pilots to fly the new plane without needing new certifications, since retraining pilots is very expensive for airlines. To achieve that goal, the airliner’s software included an array of ambitious new features, including many intended to increase safety by taking over control from the crew in certain situations. These included something called the Maneuvering Characteristics Augmentation System (MCAS), which was supposed to compensate for an excessive nose-up angle by adjusting the horizontal stabilizer to keep the aircraft from stalling— a complicated technical “hack” implemented by Boeing to avoid the larger cost involved in rewriting the program from the ground up.

The 737 MAX was a top seller right out of the gate. But what Boeing and its airline customers hadn’t realized was that the software was being asked to do things the pilots didn’t fully understand. As a result, pilots found themselves unable to interface in a timely fashion with the complex system in front of them. The ultimate result was two tragic crashes with 346 fatalities, forcing the grounding of the 737 MAX fleet and a fraud settlement that cost Boeing some $2.5 billion. Additional losses from cancelled aircraft orders, lowered stock value, and other damages have been estimated at up to $60 billion. 

These needless losses — financial and human — were caused, in large part, by small yet fatal failures of cockpit communication between people and machines. The pilots could tell that something serious was wrong, but the existing controls made it difficult for them to figure out what that was and to work with the system to correct the problem. As a result, in the words of investigative reporter Peter Robison, “the pilots were trying to retake control of the plane, so that the plane was pitching up and down over several minutes.” Based on his re-creation of what happened, Robison concludes, “it would have been terrifying for the people on the planes.”

When voice becomes a major interface in airliner cockpits, a new tool for preventing such disasters will be available. In traditional aviation, pilots receive commands like “Cleared Direct Casanova VOR” or “Intercept the ILS 3” via radio from dispatchers at air traffic control. After the pilots get this information, they must use their eyes and hands to locate and press a series of buttons to transmit the same commands to the aircraft. In a voice-driven world, that time-wasting, error-prone step will be eliminated. In the first stage of voice adoption, pilots will simply be able to say a few words without moving their eyes from the controls around them, and the plane will respond. According to Geoff Shapiro, a human factors engineer at the former Rockwell Collins Advanced Technology Center, this shift trims the time spent when entering simple navigational commands from half a minute to eight seconds — a huge improvement in circumstances when a few moments can be critical. In the second stage, once veteran pilots have recognized and accepted the power of voice, the plane will automatically follow the spoken instructions from air traffic control, merely asking the pilot to confirm them.

A voice-interface solution integrating the latest capabilities of voice-driven artificial intelligence can improve airline safety in several ways. It gives the system self-awareness and the ability to proactively communicate its state and status to pilots, thereby alerting them to problems even at moments when they might otherwise be distracted or inattentive. Using increasingly powerful voice-technology tools like automatic speech recognition and natural language understanding, it also allows the airplane’s control systems to process and act on conversational speech, making the implementation of pilot commands faster and more accurate than ever. It facilitates real-time communications linking the cockpit, air traffic control, the airline carrier, and maintenance engineers to remove inconsistencies in communication due to human indecision or misjudgment. In the near future, it may even be able to use emerging voice-tech tools such as voice biometrics and real-time sentiment analysis to determine stress levels being experienced by pilots —information that could be used to transmit emergency alerts to air traffic controllers and others on the ground.

Voice technology won’t eliminate all the traditional activities pilots are trained to perform. But in critical moments when the speed of response to messages from a control tower may spell the difference between survival and disaster, the use of a voice interface will prevent crashes and save lives. This is not a fantasy about the remote future. Today’s planes have all the electronics needed to make it possible. 

One field of aviation in which safety risks are especially intense is military flying. It’s also an arena in which voice-enabled aviation is being avidly pursued. Alabama-based Dynetics has received $12.3 million from DARPA, the Pentagon’s storied defense-technology division, to develop the use of AI in “high-intensity air conflicts.” The third phrase of the current three-phase research/implementation program involves a “realistic, manned-flight environment involving complex human-machine collaboration” — including voice communication. 

The US Air Force is not alone in pursuing this technological advantage. The next generation of the MiG-35, the highly advanced Russian fighter jet, will apparently feature a voice assistant to offer advice in high-pressure situations. Test pilot Dmitry Selivanov says, “We call her Rita, the voice communicant. Her voice remains pleasant and calm even if fire hits the engine. She does not talk all the time, she just makes recommendations if the plane approaches some restrictions. Hints are also provided during combat usage.”

Voice-controlled flying is also in development for civilian aircraft. Companies like Honeywell and Rockwell are designing voice interfaces for aviation, with an initial focus on reducing pilot workload around tedious tasks involving basic, repetitive commands like “Give me the weather at LAX and any critical weather en route.” More extensive and sophisticated use cases for voice tech in aviation are steadily emerging. Vipul Gupta is general manager of Honeywell Aerospace Avionics. He and his team are deeply focused on perfecting the technology of the voice cockpit, especially its response speed, which is a crucial safety feature. Their engineers have reduced the voice system’s average response time to 250 milliseconds, which means, in effect, that the system can react more quickly than a human pilot can.

Over time, voice-controlled aircraft systems will become commonplace in most forms of aviation. But in the short term, the most important use cases will be in general aviation, where single-pilot operators are notoriously overloaded, especially when operating in bad weather or congested areas. Having a “voice copilot” will ease those burdens and make the flying experience safer for pilot and passengers alike.

Voice-controlled aircraft are also likely to dominate the emerging field of urban air mobility, which involves the use of small aircraft for purposes ranging from cargo deliveries to sightseeing tours within city and suburban airspaces. New types of aircraft, such as electric vertical takeoff and landing aircraft (eVTOLs) are likely to dominate this domain, with the marketplace for eVTOLs expected to explode from nothing in 2022 to $1.75 billion in 2028. As this new domain of flight expands, experienced pilots will be in short supply, so the industry is now designing simplified cockpit systems, controlled by voice, that trained “operators” will be able to manage.

Vipul Gupta is bullish about the future of the voice-powered cockpit. “Eventually,” he says, “we’ll have a voice assistant where you will just sit in [the aircraft] and the passenger will say, ‘Hey, fly me there, take me there. And then the system does it.’”

As a licensed pilot with significant personal experience in the cock- pit, I suspect he will be right —eventually. As with most innovations, I believe it will take longer than the early adopters and enthusiasts believe. This is especially likely in a critical field like aviation, in which human trust issues and regulatory hurdles can take years to overcome. But the fact is that the challenges of voice-powered flight are actually simpler in many ways than those faced by other technologies, such as autonomous automobiles. For example, a plane cruising at 20,000 feet doesn’t have to deal with red lights, kids dashing into the street, or other drivers tailgating.

For this reason, I concur with the experts who say that we will have safe, effective voice-controlled planes sooner than autonomous cars. And once the technology is fully developed, the safety advantages of a system that can respond to spoken commands almost instantly in an emergency will be too powerful for the aviation industry to forgo.

This article originally appeared on Engadget at https://www.engadget.com/hitting-the-books-the-sound-of-the-future-tobias-dengel-publicaffairs-143020776.html?src=rss

Starlink's satellite cell service is set to launch in 2024, but only for SMS

The launch of Starlink's much-anticipated satellite cellular service, Direct-to-Cell, will reportedly begin rolling out for SMS in 2024, according to a newly published promotional site by the company. Eventually the system will "enable ubiquitous access to texting, calling, and browsing wherever you may be on land, lakes, or coastal waters," and connect to IoT devices through the LTE standard.

Starlink has partnered with T-Mobile on the project, which was originally announced last August at the "Coverage and Above and Beyond" event. The collaboration sees T-Mobile setting aside a bit of its 5G spectrum for use by Starlink's second-generation satellites; Starlink in turn will allow T-Mobile phones to access the satellite network giving the cell service provider "near complete coverage" of the United States. 

During the event last August, SpaceX CEO Elon Musk tweeted that "Starlink V2" would launch this year on select mobile phones, as well as in Tesla vehicles. “The important thing about this is that it means there are no dead zones anywhere in the world for your cell phone,” Musk said in a press statement at the time. “We’re incredibly excited to do this with T-Mobile.” That estimate was revised during a March panel discussion at the Satellite Conference and Exhibition 2023, when SpaceX VP of Starlink enterprise sales Jonathan Hofeller estimated testing — not commercial operation — would begin in 2023

The existing constellation of 4,265 satellites are not compatible with the new cell service so Starlink is going to have to launch a whole new series of microsats with the necessary eNodeB modem installed, over the next few years. As more satellites are launched, the adde voice and data features will become available. 

As an messaging-only satellite service, Direct-to-Cell will immediately find competition from Apple, with its Emergency SOS via Satellite feature in iOS 14, as well as Qualcomm's rival Snapdragon Satellite, which delivers texts to Android phones from orbit using the Iridium constellation. Competition is expected to be fierce in this emerging market, Lynk Global CEO Charles Miller noted during the March event, arguing that satellite cell service could potentially be the "biggest category in satellite."

This article originally appeared on Engadget at https://www.engadget.com/starlinks-satellite-cell-service-is-set-to-launch-in-2024-but-only-for-sms-215036124.html?src=rss

You can now generate AI images directly in the Google Search bar

Back in the olden days of last December, we had to go to specialized websites to have our natural language prompts transformed into generated AI art, but no longer! Google announced Thursday that users who have opted-in for its Search Generative Experience (SGE) will be able to create AI images directly from the standard Search bar.

SGE is Google’s vision for our web searching future. Rather than picking websites from a returned list, the system will synthesize a (reasonably) coherent response to the user’s natural language prompt using the same data that the list’s links led to. Thursday’s updates are a natural expansion of that experience, simply returning generated images (using the company’s Imagen text-to-picture AI) instead of generated text. Users type in a description of what they’re looking for (a Capybara cooking breakfast, in Google’s example) and, within moments, the engine will create four alternatives to pick from and refine further. Users will also be able to export their generated images to Drive or share them via email.

Google

What’s more, users will be able to generate images directly in Google Images. So, if you’re looking for (again, Google’s example) “minimalist halloween table settings” or “spooky dog house ideas,” you’ll be able to type that into the search bar and have Google generate an image based on it. What’s really cool is that you can then turn Google Lens on that generated image to search for actual, real-world products that most closely resemble what the computer hallucinated for you. 

There are, of course, a number of limitations built into the new features. You’ll have to be signed up for Google Labs and have opted-in to the SGE program to use any of this. Additionally, the new image generation functions will be available only within the US, in English-language applications and for users over the age of 18. That last requirement is a just bit odd given that Google did just go out of its way to make the program accessible to teens.

The company is also expanding its efforts to rein in the misuse of generative AI technology. Users will be blocked from creating photorealistic images of human faces. You want a photorealistic capybara cooking bacon, that’s no problem. You want a photorealistic Colonel Sanders cooking bacon, you’re going to run into issues and not just in terms of advertising canon. You’ll also be prevented from generating images of “notable” people, so I guess Colonel Sanders is out either way. 

Finally, Google is implementing the SynthID system developed by DeepMind announced last month. SythID is a visually undetectable metadata watermark that denotes a generated image as such, as well as provides information on who, or what, created it and when. The new features will be available through the Labs tab (click the flask icon) in the Google app on iOS and Android, and on Chrome desktop to select users today and expanding to more in the coming weeks.

This article originally appeared on Engadget at https://www.engadget.com/you-can-now-generate-ai-images-directly-in-the-google-search-bar-160020809.html?src=rss