Will Siri team up with ChatGPT?
PLUS: Eyes peeled for today's OpenAI event, Alibaba's compute-for-equity model
Today’s top AI Highlights:
Apple is closing a deal with OpenAI to use ChatGPT to power AI features in Apple devices
What to expect at OpenAI’s Spring Event today
Alibaba is growing cloud compute infrastructure revenue with a unique compute-for-equity model
Meta, Microsoft, and UC Berkeley’s new research for enhancing LLMs’ domain-specific performance
Use Claude to create production-ready prompts optimized for LLMs
& so much more!
Read time: 3 mins
Latest Developments 🌍
Will Siri be powered by ChatGPT? 🗣️
Apple has realized they need to bark the “generative AI” word much louder than before. The company has not trained any LLM in-house but has built powerful chips to run AI on Apple devices. With this, it is nearing a deal with OpenAI to bring ChatGPT to iPhones. The deal would see ChatGPT features integrated into the upcoming iOS 18 operating system and powering AI in more apps.
Key Highlights:
Deal with Google: Apple has also engaged in discussions with Google about potentially licensing its Gemini chatbot, although no agreement has been reached yet.
Siri by GPT-4: Rumors are flying that Siri might start using ChatGPT’s brain to answer your questions. Imagine asking Siri something and getting smarter replies.
Why it matters: If Siri gets this upgrade, it means your iPhone will not just respond faster but also smarter. You could ask for cooking tips, solve math problems, or even get relationship advice.
Stay tuned: Nothing’s official yet, but keep your eyes peeled at the WWDC 2024 on June 10.
What to Expect at OpenAI’s Event Today 🌟
OpenAI’s Spring event is scheduled for today and the social media platforms are brimming with speculations. Till last week, it was even rumored that OpenAI is planning to launch a search engine to compete with Google, however, Sam Altman has clearly negated this rumor. It’ll be all about new features and updates to ChatGPT. Let’s see what the speculations are and what we can expect today:
Connected Apps: ChatGPT will soon have the Connected Apps feature, integrating with Google Drive and Microsoft One Drive. This will let you attach documents from the cloud directly in ChatGPT, saving the unnecessary hassle of downloading the document first and uploading it again.
New Voice Assistant: ChatGPT already has a Voice Assistant that uses a transcription model (Whisper), an LLM (GPT 4), and a text-to-speech model (TTS-1). But new reports say that OpenAI has developed an advanced Voice Assistant that is much more capable than the existing one, especially at understanding the mood of the speaker. This would make up for a really good AI customer rep.
Phone Calls by ChatGPT: OpenAI might release a new feature allowing ChatGPT to make phone calls using WebRTC technology. This means users could opt-in to have real-time voice calls with ChatGPT. We would probably receive ChatGPT’s calls to get reminders for meeting or completing a project, or call us back once it is done with its research.
New Models: OpenAI might introduce three new models -
gpt4-lite: a lighter version of the current GPT-4 model to replace GPT-3.5. It might be smaller, possibly faster but still maintains some of the capabilities of GPT-4.
gpt4-auto: This could be a new model endpoint that can automatically gather data from the web and other sources. This would be different from the current capability where you have to explicitly tell GPT-4 to search the web to get the latest information.
gpt4-lite-auto: A combination of the above two features.
Alibaba’s Compute-for-Equity Model to Boost Revenue 💸
Alibaba, the Chinese e-commerce giant, is taking a novel approach to increase its revenue while also promoting China’s generative AI scene. Instead of traditional cash investments, Alibaba is leveraging its vast cloud computing infrastructure to offer computing credits to promising AI startups. This strategy allows Alibaba to secure stakes in these companies while also boosting its cloud business.
Key Highlights:
Compute-for-Equity: Alibaba is providing startups with credits to use its powerful cloud computing resources, specifically for training AI models, in exchange for equity in the companies.
Advantage: This approach addresses the scarcity of advanced computing resources in China due to US chip export restrictions. This also mirrors Microsoft’s successful investment in OpenAI by supporting local equivalents of ChatGPT and other AI applications.
Example: Alibaba led a $1 billion funding round in Moonshot AI that is building an LLM that can handle long inputs. Nearly half of Alibaba’s $800 million contribution was in the form of cloud computing credits.
The Dual Incentive: This not only allows Alibaba to secure a foothold in the rapidly growing AI sector but also drives revenue for its cloud computing division, which has experienced slowing growth in recent quarters.
Technical Research 🔬
RAFTing LLM towards More Domain-Specific Accuracy 👩⚕️
LLMs are trained on a vast and diverse dataset to make them capable of answering questions on a wide range of topics. However, many usecases require LLMs to specialize more in a specific domain. In this case, supervised fine-tuning becomes very laborious and the quality of RAG depends heavily on the quality of the document retrieval process.
Researchers at UC Berkeley, Meta AI, and Microsoft have proposed a method called Retrieval Augmented Fine-Tuning (RAFT) for LLMs to focus only on the important parts of documents in a specific domain without relying on the document retriever.
Key Highlights:
Selective Training: RAFT trains language models to differentiate between “oracle” documents that contain answers and “distractor” documents that do not, enhancing focus on relevant information.
Model and Tools: RAFT was applied on the Llama 2-7B model, trained on Microsoft AI Studio. It was chosen due to its language understanding, math skills, and ability to parse long documents.
Benchmarks and Performance: After using RAFT, Llama 2-7B was evaluated on datasets like PubMed, HotpotQA, and Gorilla, where it showed major improvements in accuracy for domain-specific Q&A.
Deployment Flexibility: The trained models can be deployed on various platforms, including GPUs or CPUs, via Microsoft AI Studio and llama.cpp, making it adaptable for different enterprise needs.
😍 Enjoying so far, share it with your friends!
Tools of the Trade ⚒️
Anthropic Console: Generate production-ready prompts in this Console using Claude-3 itself. Just describe the task, and Claude-3 will turn it into a high-quality prompt that works best with LLMs, specifically Claude-3. You can even invite team members to collaborate and use these prompts.
llm-ui: A React library for building user interfaces for LLMs. It provides features like correcting broken markdown syntax, custom component integration, and output throttling for a smoother user experience. llm-ui offers code block rendering using Shiki and a headless design allowing for complete style customization.
StartKit AI: Provides boilerplate code with OpenAI integration, Node.js API, Mongo Database, and Pinecone vector storage to quickly build AI tools and SaaS products. You can use it to create applications such as ChatGPT clones, PDF analysis tools, image generation apps, and more, with pre-built modules for common AI tasks.
Mock-My-Mockup: Upload a screenshot of a page you’re working on, and get brutally honest feedback. It’ll highlight both the positives and negatives of the page, with a little roast.
Hot Takes 🔥
Our advantage currently lies not in compute, cloud, or chips. 𝐎𝐮𝐫 𝐚𝐝𝐯𝐚𝐧𝐭𝐚𝐠𝐞 𝐢𝐬 𝐨𝐮𝐫 𝐩𝐨𝐩𝐮𝐥𝐚𝐭𝐢𝐨𝐧 𝐚𝐧𝐝 𝐭𝐡𝐞𝐢𝐫 𝐚𝐬𝐩𝐢𝐫𝐚𝐭𝐢𝐨𝐧𝐬. 𝐓𝐡𝐢𝐬 𝐢𝐬 𝐰𝐡𝐲 𝐰𝐞 𝐡𝐚𝐯𝐞 𝐭𝐨 𝐛𝐫𝐢𝐧𝐠 𝐝𝐨𝐰𝐧 𝐭𝐡𝐞 𝐜𝐨𝐬𝐭 𝐨𝐟 𝐢𝐧𝐟𝐞𝐫𝐞𝐧𝐜𝐞 𝐟𝐫𝐨𝐦 𝐑𝐬 100 𝐭𝐨 𝐑𝐬 1……𝐈𝐧𝐧𝐨𝐯𝐚𝐭𝐞 𝐟𝐫𝐮𝐠𝐚𝐥𝐥𝐲 𝐭𝐨 𝐝𝐫𝐚𝐦𝐚𝐭𝐢𝐜𝐚𝐥𝐥𝐲 𝐫𝐞𝐝𝐮𝐜𝐞 𝐭𝐡𝐞 𝐜𝐨𝐬𝐭 𝐨𝐟 𝐀𝐈. 𝐖𝐞 𝐜𝐚𝐧'𝐭 𝐝𝐞𝐥𝐢𝐯𝐞𝐫 𝐢𝐭 𝐭𝐨 𝐚 𝐛𝐢𝐥𝐥𝐢𝐨𝐧 𝐩𝐞𝐨𝐩𝐥𝐞 𝐮𝐧𝐥𝐞𝐬𝐬 𝐰𝐞 𝐜𝐚𝐧 𝐛𝐞𝐠𝐢𝐧 𝐭𝐨 𝐜𝐡𝐚𝐫𝐠𝐞 1 𝐫𝐮𝐩𝐞𝐞 𝐩𝐞𝐫 𝐭𝐫𝐚𝐧𝐬𝐚𝐜𝐭𝐢𝐨𝐧. ~
Nandan Nilekani on India’s future in AIIf OpenAI really is releasing a voice assistant tomorrow, it’s highly likely to be a true end-to-end system.
It’s the natural way forward… a somewhat-better multi-model system won’t be enough to really wow anyone.
End-to-end with low latency would be a breakthrough. ~
Matt Shumer
Meme of the Day 🤡
The OpenAI team is ready to steamroll some AI startups on Monday.
That’s all for today! See you tomorrow with more such AI-filled content.
Real-time AI Updates 🚨
⚡️ Follow me on Twitter @Saboo_Shubham for lightning-fast AI updates and never miss what’s trending!
PS: I curate this AI newsletter every day for FREE, your support is what keeps me going. If you find value in what you read, share it with your friends by clicking the share button below!