Elgato's Stream Deck Neo is 15 percent off right now https://o.aolcdn.com/images/dims?image_uri=https%3A%2F%2Fs.yimg.com%2Fos%2Fcreatr-uploaded-images%2F2024-09%2F441b2ba0-6a1a-11ef-b6ff-e0f43ccb04a6&resize=1400%2C993&client=19f2b5e49a271b2bde77&signature=e0d07d5f6a8d085646955f99b33e73c043a32e98 Folks who are looking to upgrade their livestreaming setup or become more productive on their PC or Mac may be interested in picking up an Elgato Stream Deck Neo. The desktop device allows you to set up all manner of commands and shortcuts that you can trigger with the press of a button. The Stream Deck Neo strikes a good middle ground between the Mini (which has just six buttons) and the full-size model. The Stream Deck Neo typically costs $100. However, it's currently on sale at Amazon for $85. The Stream Deck MK.2 has long been one of our recommended accessories for folks who stream gameplay on the likes of Twitch or YouTube. But the smaller Neo can do the trick too. The larger model has 15 LCD keys compared with the Neo's eight, but you can set up multiple pages of shortcuts for the Neo and navigate between them with the press of a button. Setting up commands and shortcuts through the Stream Deck PC and Mac app can be as simple as dragging and dropping actions into place. There are a ton of presets for you to get started with too. The Stream Deck Neo offers integrations with a bunch of apps, such as Zoom, Teams, PowerPoint, Excel, Word, Google Workspace, Photoshop, Adobe Creative Apps, Spotify and Music. You can use it to control smart home devices as well, including lights and speakers.  An info bar along the bottom can persistently display key information, such as the date and time. What's more, you can customize the image that appears on each LCD button to be just about whatever you want for that ultra-personal touch. Follow @EngadgetDeals on Twitter and subscribe to the Engadget Deals newsletter for the latest tech deals and buying advice. This article originally appeared on Engadget at
A creepy, sonar-like sound coming through Starliner’s speaker posed a brief mystery on the ISS https://o.aolcdn.com/images/dims?image_uri=https%3A%2F%2Fs.yimg.com%2Fos%2Fcreatr-uploaded-images%2F2024-07%2Fb4fca590-478a-11ef-87b7-b59a05f2b759&resize=1400%2C875&client=19f2b5e49a271b2bde77&signature=3f3ab644e37f7241b4ee6ff290a995efde4f22b5 Starliner is scheduled to undock from the International Space Station and make its return trip to Earth uncrewed in just a matter of days, but it apparently still has a few new mysteries left in it to throw at the team before it departs. On Saturday, astronaut Butch Wilmore alerted NASA’s Mission Control about an unexplained “strange noise” coming from a speaker in the spacecraft, which you can hear in an audio clip of the conversation shared on a NASASpaceflight forum by meteorologist Rob Dale (spotted by Ars Technica). It starts at around the 45-second mark, ringing out on a steady beat. “I don’t know what’s making it,” Wilmore said. NASA has since said that the sound has stopped in a statement to SpaceNews’ Jeff Foust on Monday, and attributed it to an audio configuration between the ISS and Starliner. It was just speaker feedback, according to the space agency. The noise was a bit of a head-scratcher over the weekend. After confirming with Mission Control on Saturday that they could hear the sound too, once Wilmore brought his mic over to the speaker, the flight controller in Houston said, “It was kind of like a pulsing noise, almost like a sonar ping.” Wilmore then let it play for about 20 seconds more before wrapping up the call. “Just to make sure I’m on the same page, this is emanating from the speaker in Starliner,” Mission Control asked, “you don’t notice anything else, any other noises, any weird configs in there?” The astronaut noted at the time that everything else seemed normal. “The space station audio system is complex, allowing multiple spacecraft and modules to be interconnected, and it is common to experience noise and feedback,” NASA said in its explanation to Foust on Monday. “The crew is asked to contact mission control when they hear sounds originating in the comm system.” The incident had no impact to the crew or Starliner’s departure schedule, it added. The Boeing spacecraft has been docked with the ISS since early June, and engineers have since had their hands full trying to get to the bottom of the issues that arose during its first crewed flight. When Starliner finally heads back to Earth on September 6, it’ll be leaving its crew — Wilmore and NASA astronaut Suni Williams — behind on the ISS, where they’ll continue to work for the next few months while they wait for a ride home from SpaceX in February 2025. Update, September 2 2024, 2:30PM ET: This story has been updated to include a statement and explanation from NASA. This article originally appeared on Engadget at
Google is rolling out more election-related safeguards in YouTube, search and AI https://o.aolcdn.com/images/dims?image_uri=https%3A%2F%2Fs.yimg.com%2Fos%2Fcreatr-uploaded-images%2F2024-08%2Fc61c5060-6700-11ef-be9f-ea3c0e76b8ec&resize=1400%2C787&client=19f2b5e49a271b2bde77&signature=d361b2a652ecf362597d5d28ff3b6a8bb4e70715 As the US speeds toward one of the most consequential elections in its 248-year history, Google is rolling out safeguards to ensure users get reliable information. In addition to the measures it announced late last year, the company said on Friday that it’s adding election-related guardrails to YouTube, Search, Google Play and AI products. YouTube will add information panels above the search results for at least some federal election candidates. The modules, likely similar to those you see when searching the web for prominent figures, will include the candidates’ basic details like their political party and a link to Google Search for more info. The company says the panels may also include a link to the person’s official website (or other channel). As Election Day (November 5) approaches, YouTube’s homepage will also show reminders on where and how to vote. Google Search will include aggregated voter registration resources from state election offices for all users. Google is sourcing that data through a partnership with Democracy Works, a nonpartisan nonprofit that works with various companies and organizations “to help voters whenever and wherever they need it.” Meanwhile, the Google Play Store will add a new badge that indicates an app is from an official government agency. The company outlines its requirements for apps that “communicate government information” in a developer help document. Approved applications that have submitted the required forms are eligible for the “official endorsement signified by a clear visual treatment on the Play Store.” As for generative AI, which can be prone to hallucinations that would make Jerry Garcia blush, Google is expanding its election-related restrictions, which were announced late last year. They’ll include disclosures for ads created or generated using AI, content labels for generated content and embedded SynthID digital watermarking for AI-made text, audio, images and video. Initially described as being for Gemini (apps and on the web), the election guardrails will apply to Search AI Overviews, YouTube AI-generated summaries for Live Chat, Gems (custom chatbots with user-created instructions) and Gemini image generation. This article originally appeared on Engadget at
This startup wants to be the iTunes of AI content licensing https://o.aolcdn.com/images/dims?image_uri=https%3A%2F%2Fs.yimg.com%2Fos%2Fcreatr-uploaded-images%2F2024-08%2Ffcdcec30-6166-11ef-bcff-393cdbb1da9f&resize=1400%2C767&client=19f2b5e49a271b2bde77&signature=3bdebc21e124b3dfcd5a54b6c2274ba5f10acc63 TollBit The 28-year-old founders of TollBit, a New York-based startup that is all of six months old, think we’re living in the “Napster days” of AI. Just like people of a certain generation downloaded digital music, companies are ripping off vast swaths of the internet without paying the rights holders. They want TollBit to be the iTunes of the AI world. “It’s kind of the Wild West right now,” Olivia Joslin, the company’s co-founder and chief operating officer, told Engadget in an interview. “We want to make it easier for AI companies to pay for the data they need.” Their idea is simple: create a marketplace that connects AI companies that need access to fresh, high-quality data to the publishers who actually spend money creating it. AI companies have, indeed, only recently started paying for (some of) the data they need from news publishers. OpenAI kicked off an arms race at the end of 2022, but it was only a year ago that the company signed the first of its many licensing deals with the Associated Press. Later that year, OpenAI announced a partnership with German publisher Axel Springer, which operates Business Insider and Politico in the US. Multiple publishers including Vox, the Financial Times, News Corp and TIME, have since signed deals with OpenAI and Google. But that still leaves countless other publishers and creators out in the cold — without the option to strike this Faustian Bargain even if they want to. This is the “long tail” of publishers that TollBit wants to target. “Powerful AI models already exist and they have already been trained,” Toshit Panigrahi, TollBit’s co-founder and CEO told Engadget. “And right now, there are thousands of applications just taking these existing models off the shelves. What they need is fresh content. But right now, there’s no infrastructure — neither for them to buy it, nor for content-makers to sell it in a way that is seamless.” Both Joslin and Panigrahi weren’t particularly knowledgeable about the media industry. But they both knew how online marketplaces and platforms operated – they were colleagues at Toast, a platform that lets restaurants manage billing and reservations. Panigrahi watched both the deals — and the lawsuits — pile up in the AI sector, then called on Joslin. Their early conversations were about RAG, which stands for Retrieval-Augmented Generation in the AI world. With RAG, AI models first look up information from specific databases (like the scrapable portions of the internet) and use that information to synthesize a response instead of simply relying on training data. Services like ChatGPT don’t know current home prices, or the latest news. Instead, they fetch that data, typically by looking at websites. That absence of fresh data is why AI chatbots are often stumped by queries about breaking news events — if they don’t scrape the latest data, they simply can’t keep up. “We thought that using content for RAG was something fundamentally different than using it for training,” said Panigrahi. TollBit By some estimations, RAG is the future of search engines. More and more, people are asking questions on the internet and expecting complete answers in return instead of a list of blue links. In just over a year, startups like Perplexity, backed by Jess Bezos and NVIDIA among others, have burst onto the scene with ambitions of taking on Google. Even OpenAI has plans to someday let ChatGPT become your search engine. In response, Google has sprung into action — it now culls relevant information from search results and presents it as a coherent answer at the top of the results page, a feature it calls AI Overviews. (It doesn’t always work well, but is seemingly here to stay). The rise of RAG-based search engines has publishers shaking in their boots. After all, who would make money if AI reads the internet for us? After Google rolled out AI Overviews earlier this year, at least one report estimated that publishers would lose more than $2 billion in ad revenue because fewer people would have a reason to visit their websites. “AI companies need continuous access to high quality content and data too,” said Joslin, “but if you don’t figure out some economic model here, there will be no incentive for anyone to create content, and that’ll be the end of AI applications too.” Instead of cutting one-off checks, TollBit’s model aims to compensate publishers on an ongoing basis. Hypothetically, if someone’s content was used in a thousand AI-generated answers, they would get paid a thousand times at a price that they set and which they can change on the fly. Each time an AI company accesses fresh data from a publisher through TollBit, it can pay a small fee set by the publisher that Panigrahi and Joslin think should be roughly equivalent to whatever a traditional page view would have made the publisher. And the platform can also block AI companies who haven’t signed up from accessing publishers’ data. So far, the founders claim to have onboarded a hundred publishers and are in pilots with three AI companies since TollBit launched in February. They refused to reveal which publishers or AI companies had signed on so far, citing confidentiality clauses, but did not deny speaking with OpenAI, Anthropic, Google and Meta. So far, they say that no money has changed hands between AI companies and publishers on their platform. TollBit Until that happens, their model is still a giant hypothetical — although one that investors have so far poured $7 million into. TollBit’s investors include Sunflower Capital, Lerer Hippeau, Operator Collective, AIX and Liquid 2 Ventures, and more investors are currently “pounding down their door,” Joslin claimed. In April, TollBit also brought on Campbell Brown as a senior adviser, a former television anchor who previously acted as Meta’s head of news partnerships for the better part of a decade. In spite of some high-profile lawsuits, AI companies are still scraping the internet for free and largely getting away with it. Why would they have any incentive to actually pay publishers for this data? There are three big reasons, the founders say: more websites are taking steps to prevent their content from being scraped ever since generative AI went mainstream, which means that scraping the web is getting harder and more expensive; no one wants to deal with ongoing copyright lawsuits; and, crucially, being able to easily pay for content on an as-needed basis lets AI companies tap into smaller and more niche publications because it isn’t possible to strike individual licensing deals with every single website. Joslin also pointed out that multiple TollBit investors have also invested in AI companies which they worry might face litigation for using content without permission. Getting AI companies to pay for content could provide a recurring revenue stream for not just large publishers but to potentially anyone who publishes anything online. Last month, Perplexity — which was accused of illegally scraping content from Forbes, Wired and Condé Nast — launched a Publishers’ Program under which it plans to share a cut of any revenue it earns with publishers if it uses their content to generate answers with AI. The success of the program, however, hinges on how much money Perplexity makes when it introduces ads in the app later this year. Like Tollbit, it's another complete hypothetical. “Our thesis with TollBit is that if you lose a page view today, you should be compensated for it immediately rather than a few years after when a tech company figures out its ads program,” said Panigrahi about Perplexity’s initiative. Despite all the existing licensing deals and technical advances, AI-powered chatbots still make for terrible news sources. They still make up facts and confidently conjure up entire links to stories that don’t actually exist. But technology companies are now stuffing AI chatbots in every crevice they can, which means that many people will still get their news from one of these products in the not-so-distant future. A more cynical take on TollBit’s premise is that the startup is effectively offering hush money to publishers whose work is more likely than not to be sausaged into misinformation. Its founders, naturally, don’t agree with the characterization. “We are careful about the AI partners we onboard,” Panigrahi said. “These companies are very mindful about the quality of input material and correctness of responses. We’re seeing that paying for content – even nominal amounts – creates incentive to respect the raw inputs into their systems instead of treating it as a free, replaceable commodity.” This article originally appeared on Engadget at
DeepMind workers urge Google to drop military contracts https://o.aolcdn.com/images/dims?image_uri=https%3A%2F%2Fs.yimg.com%2Fos%2Fcreatr-uploaded-images%2F2024-08%2F8c000a00-60b3-11ef-bf3f-cccfd3f9261d&resize=1400%2C933&client=19f2b5e49a271b2bde77&signature=1912758b0d5445b546cc08dfa6c584731218303d Google DeepMind workers have signed a letter calling on the company to drop contracts with military organizations, according to a report by Time. The document was drafted on May 16 of this year. Around 200 people signed the document, which amounts to five percent of the total headcount of DeepMind.  For the uninitiated, DeepMind is one of Google’s AI divisions and the letter states that adopting military contracts runs afoul of the company’s own AI rules. The letter was sent out as internal concerns began circulating within the AI lab that the tech was allegedly being sold to military organizations via cloud contracts. According to Time, Google’s contracts with the United States military and the Israeli military allow access to services via the cloud, and this reportedly includes AI technology developed by DeepMind. The letter doesn’t linger on any specific military organization, with workers emphasizing that it’s “not about the geopolitics of any particular conflict.”  Reporting since 2021 has slowly revealed the scope of tech supplied by Google (and Amazon) to the Israeli government via a partnership known as Project Nimbus. This is far from the first instance of Google employees openly protesting their work being used to support politically fraught military aims — the company fired dozens of staffers who spoke out against Project Nimbus earlier this year. “Any involvement with military and weapon manufacturing impacts our position as leaders in ethical and responsible AI, and goes against our mission statement and stated AI principles,” the DeepMind letter says. It’s worth noting that Google’s slogan used to be “don’t be evil.” The letter goes on to ask DeepMind’s leaders to deny military users access to its AI technology and to set up a new in-house governance body to prevent the tech from being used by future militaries. According to four unnamed employees, Google has yet to offer a tangible response to the letter. “We have received no meaningful response from leadership,” one said, “and we are growing increasingly frustrated.” Google did respond to Time’s reporting, saying that it complies with its AI principles. The company says that the contract with the Israeli government “is not directed at highly sensitive, classified or military workloads relevant to weapons or intelligence services.” However, its partnership with the Israeli government has fallen under plenty of scrutiny in recent months.  Google purchased DeepMind back in 2014, but under the promise that its AI technology would never be used for military or surveillance purposes. For many years, DeepMind was allowed to operate with a good amount of independence from its parent company, but the burgeoning AI race looks to have changed that. The lab's leaders spent years seeking greater autonomy from Google, but were rebuffed in 2021. This article originally appeared on Engadget at