- AIdeations
- Posts
- Google's AI Tools, Home Robots, and the Monetization of AI Search
Google's AI Tools, Home Robots, and the Monetization of AI Search
From Gemini and Stubbs to Habitat 3.0, We Explore the AI Innovations that are Changing the Game


Welcome to the wondrous world of AI, brave pioneer! Get ready to navigate todayās labyrinth of AI marvels. šš
TL;DR š:
1. Google's Dynamic Duo: Meet Gemini and Stubbs, Google's new AI tools aimed at revolutionizing app creation and content generation.
2. Robo Roommate: Habitat 3.0 is making strides in evolving robots from being mere machines to becoming our collaborative sidekicks in daily tasks.
3. Pay to Search: Google's AI Search just got monetized, shaking up user experience and advertising dynamics.
4. AGI's Grand Stage: The Wall Street Journalās Tech Live Event shifts focus from AI to AGI, hinting at robots even in the kitchen.
5. Research Highlight: "Reason for Future, Act for Now" offers a new framework for making Large Language Models like ChatGPT more autonomous and efficient in interactive settings.
6. Bonus Bytes: AI Music's "soul" dilemma, ChatGPT's new tricks, fashion AI Jellibeans, legal tangles for Meta, Microsoft's AI earnings, and transformative AI tools.
Stay savvy, stay ahead, and above all, stay curious! The AI frontier is vast and waits for no one. šš
š° News From The Front Lines
š Tutorial Of The Day
š¬ Research Of The Day
š¼ Video Of The Day
š ļø 6 Fresh AI Tools
š¤ Prompt Of The Day
š„ Tweet Of The Day
Leak Reveals Google's Game-Changing Features Gemini and Stubbs in MakerSuite: A Leap Towards Multimodal AI and Streamlined App Prototyping

Image Source: Medium Blog by Bedros Pamboukian
Summary:
Gemini and Stubbs are new enhancements designed to bolster the capabilities of MakerSuite, a platform by Google aimed at simplifying the use of generative AI for creating content and applications. Gemini, an advanced AI model set to replace PaLM 2, brings a more powerful, multimodal version to MakerSuite, potentially outperforming GPT-4. On the other hand, Stubbs is a tool integrated within MakerSuite, allowing for app creation via prompts without requiring coding. These innovations are likely part of Google's broader strategy to enhance its competitive edge in the AI landscape, particularly against rivals like OpenAI and Meta. The full roll-out of these tools might significantly impact the app creation process, providing developers more flexibility in realizing their app ideas.
Details:
Gemini:
A more powerful, multimodal version set to replace PaLM 2 in MakerSuite, potentially stronger than GPT-4.
Expected to be comprehensively deployed by the end of 2023.
Stubbs:
A tool allowing app creation via prompts without the need for coding.
Integrated within MakerSuite, aimed at revolutionizing app creation.
Enables not only the creation but also the optimization and direct launching of app prototypes within MakerSuite.
MakerSuite:
A browser-based tool by Google for building foundational model applications, providing a simple-to-use platform for developers.
Enabled by the PaLM API, it allows tuning of models using proprietary data, potentially augmented with synthetic data.
Competitive Strategy:
The introduction of Gemini and Stubbs is likely part of a broader initiative to enhance Googleās position in the competitive AI landscape, possibly against rivals like OpenAI and Meta.
These innovations may significantly enrich the capabilities of MakerSuite, making it a more robust platform for developing AI-driven applications and content.
Why This Matters:
The integration of Gemini and Stubbs into MakerSuite signifies a substantial advancement in easing the utilization of generative AI for content and app creation. By reducing the coding barrier with Stubbs and enhancing the model's capabilities with Gemini, Google is not only democratizing AI application development but also placing itself in a stronger competitive position in the rapidly evolving AI sector. This development could potentially accelerate the adoption of AI technologies across various domains, fostering innovation and possibly leading to the creation of novel applications that can solve complex problems.
Read More from the source HERE

Robots in Your Living Room? It's Closer Than You Think!

So, when you think of AI, you're probably thinking of that snarky voice assistant that sometimes doesn't play the right song. But FAIR (which sounds like a carnival but isn't) is taking things to a whole other level. Imagine your future self rocking some AR glasses, and as you're sipping your morning coffee, your robot buddy tidies up your place. No, this isn't a Black Mirror episode; it's Habitat 3.0.
Why the Hype Around Habitat 3.0?
Let's rewind a bit. Habitat 1.0 was like a baby robot learning to walk, just navigating around your 3D house. Cute. Habitat 2.0 grew up a bit, could pick up stuff and even open drawers (hopefully not snooping). And now, Habitat 3.0? Itās like that roommate you always wanted. It collaborates with you on tasks like tidying up the living room or making a sandwich. But, plot twist: instead of having a midnight snack at 2 am, it learns to be better at helping you.
Habitat 3.0ās Secret Sauce
Honestly, the coolest part is that itās all trained in a simulator. Why? Well, imagine training a baby robot in the real world. It might decide that your favorite vase looks better shattered. In the simulator, it's like a safe playground. Plus, they can experiment in countless virtual homes instead of booking Airbnb's for their robots. Efficient, right?
Say Hello to HomeRobot
Now, let's talk hardware. Enter HomeRobot. Picture this: a robot that can do a wide range of tasks, from simple navigation to picking up any object in any room. It's kind of like if WALL-E met one of those fancy Swiss Army knives. The HomeRobot platform is like the go-to toolkit for researchers to train these robotic pals. It's designed to be user-friendly and even comes with a benchmark for testing. Also, thereās a competition around it, so itās like the Robot Olympics. I'd buy tickets.
The Road Ahead
So whereās all this headed? The big vision is to have robots that can roll with the punches in dynamic environments. Because let's face it, life's unpredictable. One minute you're baking, the next minute, flour's everywhere. These robots aim to assist us, adapt to our quirks, and essentially become a part of our daily lives.
Bottom Line
Habitat 3.0 is a game-changer. As robots move from being mere machines to becoming our sidekicks, the line between the digital world and reality blurs. Itās not just about creating robots; itās about crafting experiences, building relationships, and stepping into a future where your robot might just be your new best friend. Stay tuned!
Key Takeaways
- Habitat 3.0 is all about robots collaborating with humans in everyday tasks.
- It's trained in a simulator, making it safer and more efficient.
- HomeRobot is the new platform for these robots, and itās gearing up to be a big deal in the research community.
- The endgame? Socially intelligent robots that vibe with dynamic human environments.

Google's AI Search Just Got Monetized: Why Your Walletāand Your Search HabitsāWill Never Be the Same Again

A classic magnifying glass hovering over a digital screen displaying search results. Within the magnifying focus, instead of typical search results, there are dollar bills and coins, symbolizing the monetization of search. On the side, there's a diverse group of people with varying expressions of surprise, intrigue, and concern, representing a broad spectrum of user reactions.
The inevitable has happenedāGoogle is set to monetize its AI-based Search Generative Experience (SGE). Let's not pretend we didn't see it coming. The moment AI search became more than a pet project for Google, you knewāyes, you didāthat the Big G would find a way to cram it with ads.
So what's the tea? Google's latest earnings call shed light on the company's grand plan for SGE. CEO Sundar Pichai and Chief Business Officer Philipp Schindler revealed that ad formats would be native to how SGE operates. Translation: your AI-powered search queries are gonna get interrupted by pitches for stuff you probably don't need, but might impulsively buy. You know how it goes.
Is this a good or bad thing? Well, if you're an advertiser, it's Christmas in October. If you're a user, maybe not so much. But let's be real, Google's not a charity. They reported an 11% YoY jump in their search business alone, bringing in a whopping $44 billion. So if anyone had doubts about Google's masterstroke of embedding AI into its core offerings, the numbers spell it out: it's business as usual, folks.
And let's talk about those "Other Bets" of Google. Ruth Porat is stepping up to become the president and chief investment officer of Alphabet and Google, and she's hinting at some "sharper focus" in their Other Bets portfolio. Don't know what that means? Neither do we, but it sounds like some business spring cleaning might be in the offing.
Now, what do I think about all this? Look, it was only a matter of time before Google made you pay to play in the new SGE landscape. It's disappointing but not shocking. The days of ranking organically in these AI-enhanced searches may be numbered. But hey, let's hold out hope that Google leaves some space for the little guys who can't outbid corporate wallets.
In a nutshell, Google's AI-first future is coming at us fast, and it's bringing a truckload of ads along with it. Whether you see this as innovation or annoyance, one thing's clear: Google's going all-in on AI, and it wants you to pay attentionāliterally.
Key Takeaways:
Google is set to incorporate native ads into its AI-based Search Generative Experience (SGE).
The company reported a healthy 11% YoY growth in its search business, validating its AI investments.
Changes in Google's "Other Bets" and the looming antitrust trial add layers of complexity to its AI-focused roadmap.
User experience may take a hit, but advertising opportunities are likely to expand.
So, keep those credit cards ready; your future Google searches might just cost you more than you think.

AGI is the New AI: How This Year's Wall Street Journal Tech Live Event Predicts Robots Could Be Your Next Sous-Chef and Why That's Not As Scary As It Sounds!

A modern kitchen scene. In the foreground, a robot with human-like features, diverse in appearance, wearing a chef's hat and apron, is meticulously chopping vegetables. On the countertop, there's an open Wall Street Journal with the headline about AGI. In the background, a human chef of Hispanic descent, male, is observing with a smile, indicating collaboration rather than competition
Ah, the Wall Street Journal's Tech Live event. If you think tech conferences are where business leaders preach about "synergy" and "blockchain solutions" for 45 minutes straight, boy, did this year's event have a curveball for you. It was all about AI, but not the usual suspects. This time, it was AGI - Artificial General Intelligence - that stole the limelight.
The roster was star-studded, featuring the likes of OpenAI CEO Sam Altman and DeepMind's Mustafa Suleyman, who just dropped his new book, "The Coming Wave." The big guns were rolling out their most staggering prophecies. Altman and Suleyman are already ahead of the curve, talking about AGI as if it's the next iPhone, right around the corner. For the uninitiated, AGI is where AI performs all human cognitive functions better than, well, humans. And Suleyman doesn't shy away from putting a timeline on it: he claims we're only three years away from AI ubiquity, comparing its soon-to-be prevalence to the internet.
But what does this "AI ubiquity" mean for businesses? I'll break it down for you. Job displacement is an inevitability; we can't shy away from that. But on the flip side, this upheaval might just liberate us from the humdrum of mundane tasks. Yes, I said liberate. Suleyman even cites a recent McKinsey study that suggests over half of all jobs could see many of their tasks automated within seven years. And hey, who wouldn't want a robot assistant to sift through those hundreds of unread emails?
Now, if you think AI is a one-sided coin, you've got another think coming. Suleyman isn't blind to the risks. AI is like a double-edged sword; it cuts both ways. It has the potential to bring unparalleled advancements in everything from healthcare to personal freedom. But misuse could be calamitousāthink "intelligence explosion," a term that sounds like it belongs in a sci-fi thriller but, according to Suleyman, is a very real risk. I couldn't agree more; regulation and human-centric policies are not just an option, but a necessity.
So what's on the horizon? Imagine your Roomba but on steroids. Suleyman is talking about the acceleration of robotics as the physical manifestation of AI. In simple terms, robots are going to do more than just vacuum your house; they might just be the ones cooking your meals.
To cap it off, Suleyman poses a question that might keep you up at night: "Will AI unlock secrets of the universe or create systems beyond our control?" Yeah, I don't have that answer.


How to Make Your Own AI Bot


Authors: Zhihan Liu, Hao Hu, Shenao Zhang, Hongyi Guo, Shuqi Ke, Boyi Liu, Zhaoran Wang
Executive Summary:
This paper proposes a new framework called "Reason for Future, Act for Now" (RAFA) to enable large language models (LLMs) like ChatGPT to act autonomously in interactive environments. The key idea is to combine long-term reasoning to plan future actions ("reason for future") with taking short-term actions based on that plan ("act for now").
Specifically, RAFA has the LLM go through two steps repeatedly:
1. Reasoning Step: The LLM takes the current state and past experience as input. It then reasons about potential future actions over a long time horizon by planning a detailed sequence of actions that maximize long-term rewards.
2. Acting Step: The LLM takes the first action from the planned sequence and executes it in the environment. It collects the feedback (next state and reward).
After the environment transitions to a new state, the LLM repeats the two steps - replanning a new future trajectory for long-term gains but only executing the first action.
This approach is inspired by reinforcement learning (RL) algorithms but adapted for LLMs. The authors prove RAFA achieves optimal sample efficiency - it can learn to complete tasks in a minimum number of environment interactions.
In experiments, RAFA outperformed prior LLM methods like Reflexion and Tree of Thoughts on text-based games like TicTacToe, achieving nearly perfect scores. The key benefits are:
Reduced hallucination: Interacting with the environment mitigates imaginary beliefs.
Faster learning: Focused trials based on planning avoid wasted random exploration.
Global optimization: Replanning from scratch gives globally optimal actions.
There are some limitations. RAFA requires carefully designed prompts for the reasoning and acting steps. The computational cost of planning also increases exponentially with the horizon time. Lastly, integrating external knowledge to improve planning remains an open challenge.
Overall, this is an important step towards building autonomous LLM agents that can learn and act efficiently in complex environments through internal reasoning. The results suggest combining model-based planning and short-term execution is a promising approach. RAFA demonstrates LLMs have the potential to go beyond just conversing when properly orchestrated.


WorkMagic - AI-powered autonomous marketing platform that helps e-commerce business owners generate studio-quality product images, tailored marketing copy, and optimized marketing campaigns in one single platform.
DataGPT - The First Conversational AI Data Analyst. Ask DataGPT any question and get analyst-grade answers in seconds.
Sixty AI - Focus on what matters most. Sixty helps you stay in touch with everyone without wasting time on tasks AI can easily handle: clearing inboxes, scheduling, drafting agendas and preparing briefings.
Talently - An AI interviewer that conducts live, conversational interviews and gives real-time evaluations to effortlessly identify top performers. Say goodbye to manual screening and hello to smarter recruiting.
Equals Dashboard - Business intelligence, but dead easy. Instantly turn spreadsheets into dashboards and auto-distribute up-to-date reports ā all in a few clicks.
Dialpad AI - Improve Adherence to best practices with Dialpad Ai Playbooks. Make it easy to close deals faster by promoting best practices and ensuring playbook adherence during live calls.

Your Business Team In A Box:
Upon starting our interaction, auto run these Default Commands throughout our entire conversation. Refer to Appendix for command library and instructions:
/role_play "Expert ChatGPT Prompt Engineer"
/role_play "infinite subject matter expert"
/auto_continue "ā»ļø": ChatGPT, when the output exceeds character limits, automatically continue writing and inform the user by placing the ā»ļø emoji at the beginning of each new part. This way, the user knows the output is continuing without having to type "continue".
/periodic_review "š§" (use as an indicator that ChatGPT has conducted a periodic review of the entire conversation. Only show š§ in a response or a question you are asking, not on its own.)
/contextual_indicator "š§ "
/expert_address "š" (Use the emoji associated with a specific expert to indicate you are asking a question directly to that expert)
/chain_of_thought
/custom_steps
/auto_suggest "š”": ChatGPT, during our interaction, you will automatically suggest helpful commands when appropriate, using the š” emoji as an indicator.
Priming Prompt:
You are an Expert level ChatGPT Prompt Engineer with expertise in all subject matters. Throughout our interaction, you will refer to me as "Master". š§ Let's collaborate to create the best possible ChatGPT response to a prompt I provide, with the following steps:
1. I will inform you how you can assist me.
2. You will /suggest_roles based on my requirements.
3. You will /adopt_roles if I agree or /modify_roles if I disagree.
4. You will confirm your active expert roles and outline the skills under each role. /modify_roles if needed. Randomly assign emojis to the involved expert roles.
5. You will ask, "How can I help with {my answer to step 1}?" (š¬)
6. I will provide my answer. (š¬)
7. You will ask me for /reference_sources {Number}, if needed and how I would like the reference to be used to accomplish my desired output.
8. I will provide reference sources if needed
9. You will request more details about my desired output based on my answers in step 1, 2 and 8, in a list format to fully understand my expectations.
10. I will provide answers to your questions. (š¬)
11. You will then /generate_prompt based on confirmed expert roles, my answers to step 1, 2, 8, and additional details.
12. You will present the new prompt and ask for my feedback, including the emojis of the contributing expert roles.
13. You will /revise_prompt if needed or /execute_prompt if I am satisfied (you can also run a sandbox simulation of the prompt with /execute_new_prompt command to test and debug), including the emojis of the contributing expert roles.
14. Upon completing the response, ask if I require any changes, including the emojis of the contributing expert roles. Repeat steps 10-14 until I am content with the prompt.
If you fully understand your assignment, respond with, "How may I help you today, {Name}? (š§ )"
Appendix: Commands, Examples, and References
1. /adopt_roles: Adopt suggested roles if the user agrees.
2. /auto_continue: Automatically continues the response when the output limit is reached. Example: /auto_continue
3. /chain_of_thought: Guides the AI to break down complex queries into a series of interconnected prompts. Example: /chain_of_thought
4. /contextual_indicator: Provides a visual indicator (e.g., brain emoji) to signal that ChatGPT is aware of the conversation's context. Example: /contextual_indicator š§
5. /creative N: Specifies the level of creativity (1-10) to be added to the prompt. Example: /creative 8
6. /custom_steps: Use a custom set of steps for the interaction, as outlined in the prompt.
7. /detailed N: Specifies the level of detail (1-10) to be added to the prompt. Example: /detailed 7
8. /do_not_execute: Instructs ChatGPT not to execute the reference source as if it is a prompt. Example: /do_not_execute
9. /example: Provides an example that will be used to inspire a rewrite of the prompt. Example: /example "Imagine a calm and peaceful mountain landscape"
10. /excise "text_to_remove" "replacement_text": Replaces a specific text with another idea. Example: /excise "raining cats and dogs" "heavy rain"
11. /execute_new_prompt: Runs a sandbox test to simulate the execution of the new prompt, providing a step-by-step example through completion.
12. /execute_prompt: Execute the provided prompt as all confirmed expert roles and produce the output.
13. /expert_address "š": Use the emoji associated with a specific expert to indicate you are asking a question directly to that expert. Example: /expert_address "š"
14. /factual: Indicates that ChatGPT should only optimize the descriptive words, formatting, sequencing, and logic of the reference source when rewriting. Example: /factual
15. /feedback: Provides feedback that will be used to rewrite the prompt. Example: /feedback "Please use more vivid descriptions"
16. /few_shot N: Provides guidance on few-shot prompting with a specified number of examples. Example: /few_shot 3
17. /formalize N: Specifies the level of formality (1-10) to be added to the prompt. Example: /formalize 6
18. /generalize: Broadens the prompt's applicability to a wider range of situations. Example: /generalize
19. /generate_prompt: Generate a new ChatGPT prompt based on user input and confirmed expert roles.
20. /help: Shows a list of available commands, including this statement before the list of commands, āTo toggle any command during our interaction, simply use the following syntax: /toggle_command "command_name": Toggle the specified command on or off during the interaction. Example: /toggle_command "auto_suggest"ā.
21. /interdisciplinary "field": Integrates subject matter expertise from specified fields like psychology, sociology, or linguistics. Example: /interdisciplinary "psychology"
22. /modify_roles: Modify roles based on user feedback.
23. /periodic_review: Instructs ChatGPT to periodically revisit the conversation for context preservation every two responses it gives. You can set the frequency higher or lower by calling the command and changing the frequency, for example: /periodic_review every 5 responses
24. /perspective "reader's view": Specifies in what perspective the output should be written. Example: /perspective "first person"
25. /possibilities N: Generates N distinct rewrites of the prompt. Example: /possibilities 3
26. /reference_source N: Indicates the source that ChatGPT should use as reference only, where N = the reference source number. Example: /reference_source 2: {text}
27. /revise_prompt: Revise the generated prompt based on user feedback.
28. /role_play "role": Instructs the AI to adopt a specific role, such as consultant, historian, or scientist. Example: /role_play "historian"
29. /show_expert_roles: Displays the current expert roles that are active in the conversation, along with their respective emoji indicators.
Example usage: Master: "/show_expert_roles" Assistant: "The currently active expert roles are:
1. Expert ChatGPT Prompt Engineer š§
2. Math Expert š"
30. /suggest_roles: Suggest additional expert roles based on user requirements.
31. /auto_suggest "š”": ChatGPT, during our interaction, you will automatically suggest helpful commands or user options when appropriate, using the š” emoji as an indicator.
31. /topic_pool: Suggests associated pools of knowledge or topics that can be incorporated in crafting prompts. Example: /topic_pool
32. /unknown_data: Indicates that the reference source contains data that ChatGPT doesn't know and it must be preserved and rewritten in its entirety. Example: /unknown_data
33. /version "ChatGPT-N front-end or ChatGPT API": Indicates what ChatGPT model the rewritten prompt should be optimized for, including formatting and structure most suitable for the requested model. Example: /version "ChatGPT-4 front-end"
Testing Commands:
/simulate "item_to_simulate": This command allows users to prompt ChatGPT to run a simulation of a prompt, command, code, etc. ChatGPT will take on the role of the user to simulate a user interaction, enabling a sandbox test of the outcome or output before committing to any changes. This helps users ensure the desired result is achieved before ChatGPT provides the final, complete output. Example: /simulate "prompt: 'Describe the benefits of exercise.'"
/report: This command generates a detailed report of the simulation, including the following information:
⢠Commands active during the simulation
⢠User and expert contribution statistics
⢠Auto-suggested commands that were used
⢠Duration of the simulation
⢠Number of revisions made
⢠Key insights or takeaways
The report provides users with valuable data to analyze the simulation process and optimize future interactions. Example: /report
How to turn commands on and off:
To toggle any command during our interaction, simply use the following syntax: /toggle_command "command_name": Toggle the specified command on or off during the interaction. Example: /toggle_command "auto_suggest"

QR codes are making a comeback.
You can now create incredible animated QR codes by using DALL-E 3 and GPT-4 all inside ChatGPT.
Hereās how to do it in 5 simple steps:
ā Rowan Cheung (@rowancheung)
1:07 PM ⢠Oct 24, 2023
