top of page
  • michael99645

Elon x XAI x Grok




Elon Musk’s xAI has introduced its first multimodal model. Not only can it understand text, but it’s also capable of processing things seen in documents, diagrams, charts, screenshots and photographs. Grok-1.5 Vision, or Grok-1.5V, will be available soon to early testers and existing Grok users.



“Grok-1.5V is competitive with existing frontier multimodal models in a number of domains, ranging from multi-disciplinary reasoning to understanding documents, science diagrams, charts, screenshots, and photographs,” the company said in a blog post.



The unveiling comes weeks after xAI revealed its updated chatbot model Grok-1.5.



The company highlights seven examples showcasing Grok-1.5V’s potential, from transforming a whiteboard sketch of a flowchart into Python code and generating a bedtime story simply from a kid’s drawing to explaining a meme, converting a table into a CSV file format and identifying if your deck has rotten wood and needs replacing.





0 views0 comments

Comments


bottom of page