Search This Blog

Powered by Blogger.

Blog Archive

Labels

About Me

Showing posts with label Google AI model. Show all posts

Google’s AI Virtual Try-On Tool Redefines Online Shopping Experience

 

At the latest Google I/O developers conference, the tech giant introduced an unexpected innovation in online shopping: an AI-powered virtual try-on tool. This new feature lets users upload a photo of themselves and see how clothing items would appear on their body. By merging the image of the user with that of the garment, Google’s custom-built image generation model creates a realistic simulation of the outfit on the individual. 

While the concept seems simple, the underlying AI technology is advanced. In a live demonstration, the tool appeared to function seamlessly. The feature is now available in the United States and is part of Google’s broader efforts to enhance the online shopping experience through AI integration. It’s particularly useful for people who often struggle to visualize how clothing will look on their body compared to how it appears on models.  

However, the rollout of this tool raised valid questions about user privacy. AI systems that involve personal images often come with concerns over data usage. Addressing these worries, a Google representative clarified that uploaded photos are used exclusively for the try-on experience. The images are not stored for AI training, are not shared with other services or third parties, and users can delete or update their photos at any time. This level of privacy protection is notable in an industry where user data is typically leveraged to improve algorithms. 

Given Google’s ongoing development of AI-driven tools, some expected the company to utilize this photo data for model training. Instead, the commitment to user privacy in this case suggests a more responsible approach. Virtual fitting technology isn’t entirely new. Retail and tech companies have been exploring similar ideas for years. Amazon, for instance, has experimented with AI tools in its fashion division. Google, however, claims its new tool offers a more in-depth understanding of diverse body types. 

During the presentation, Vidhya Srinivasan, Google’s VP of ads and commerce, emphasized the system’s goal of accommodating different shapes and sizes more effectively. Past AI image tools have faced criticism for lacking diversity and realism. It’s unclear whether Google’s new tool will be more reliable across the board. Nevertheless, their assurance that user images won’t be used to train models helps build trust. 

Although the virtual preview may not always perfectly reflect real-life appearances, this development points to a promising direction for AI in retail. If successful, it could improve customer satisfaction, reduce returns, and make online shopping a more personalized experience.

Google Admits Editing of AI Viral Video for Enhanced Presentation

 

jA widely-viewed video showcasing the purported real-time interaction capabilities of Google's artificial intelligence model, Gemini, has been revealed to have been manipulated for demonstration purposes. The video, garnering 1.6 million views on YouTube, depicts seamless back-and-forth interactions between the AI and a human, responding to spoken-word prompts and visual stimuli.

Google, in the video's description, acknowledged that the presentation was not entirely genuine, as it had accelerated response times for the demonstration. Subsequently, in a blog post released concurrently with the video, Google disclosed the actual process behind its creation.

Contrary to the apparent spontaneity of the AI's responses to voice and video prompts, it was disclosed that the AI was, in reality, prompted using static image frames from the footage and text-based prompts. Google confirmed this revelation to the BBC, clarifying that the video was generated by presenting the AI with questions through text prompts and static images.

A Google spokesperson stated, "Our Hands on with Gemini demo video shows real prompts and outputs from Gemini. We made it to showcase the range of Gemini’s capabilities and to inspire developers."

In specific instances within the video, such as identifying materials of objects and tracking a hidden ball in a cups and balls magic routine, the AI's seemingly impressive responses were based on still images rather than live interactions. Google explained that the AI was provided with static images representing the challenges presented in the video, allowing it to showcase Gemini's capabilities across various scenarios.

This clarification from Google emphasizes the staged nature of the video, intended primarily as a demonstration to highlight the potential of Gemini's capabilities rather than a genuine portrayal of real-time AI responses to dynamic stimuli.