Gemini 2.0 is out, and there's a ton of interesting stuff about it. From my testing it looks like Gemini 2.0 Flash may be the best currently available multi-modal model - I upgraded my LLM plugin to support that here: https://github.com/simonw/llm-gemini/releases/tag/0.7
Gemini 2.0 announcement: https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/
Gemini 2.0 announcement: https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/
Comments
Output here: https://gist.github.com/simonw/32172b6f8bcf8e55e489f10979f8f085
https://aistudio.google.com/live only worked for me in Chrome: point your webcam at anything and have a live audio conversation about what the model can "see"
I am curious: have you thought or written about how you parse the ethical issues related to this technology?
I mean if a particular tech is made unethically, how / where do you draw the line (for yourself) between helping people understand it and becoming complicit in its harm?