sisyphean

joined 1 year ago
MODERATOR OF
 

Intelligence explosion arguments don’t require Platonism. They just require intelligence to exist in the normal fuzzy way that all concepts exist.

[–] sisyphean@programming.dev 1 points 1 year ago

It would summarize the link. Unfortunately that’s an edge case where the bot doesn’t do what you mean.

1
submitted 1 year ago* (last edited 1 year ago) by sisyphean@programming.dev to c/auai@programming.dev
 

At OpenAI, protecting user data is fundamental to our mission. We do not train our models on inputs and outputs through our API.

 

We’re rolling out custom instructions to give you more control over how ChatGPT responds. Set your preferences, and ChatGPT will keep them in mind for all future conversations.

@AutoTLDR

 

GPT-3.5 and GPT-4 are the two most widely used large language model (LLM) services. However, when and how these models are updated over time is opaque. Here, we evaluate the March 2023 and June 2023 versions of GPT-3.5 and GPT-4 on four diverse tasks: 1) solving math problems, 2) answering sensitive/dangerous questions, 3) generating code and 4) visual reasoning. We find that the performance and behavior of both GPT-3.5 and GPT-4 can vary greatly over time. For example, GPT-4 (March 2023) was very good at identifying prime numbers (accuracy 97.6%) but GPT-4 (June 2023) was very poor on these same questions (accuracy 2.4%). Interestingly GPT-3.5 (June 2023) was much better than GPT-3.5 (March 2023) in this task. GPT-4 was less willing to answer sensitive questions in June than in March, and both GPT-4 and GPT-3.5 had more formatting mistakes in code generation in June than in March. Overall, our findings shows that the behavior of the “same” LLM service can change substantially in a relatively short amount of time, highlighting the need for continuous monitoring of LLM quality.

 

Introducing Llama 2 - The next generation of our open source large language model. Llama 2 is available for free for research and commercial use.

This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters.

@AutoTLDR

 

16 Mar, 2023

Kagi Search is pleased to announce the introduction of three AI features into our product offering.

We’d like to discuss how we see AI’s role in search, what are the challenges and our AI integration philosophy. Finally, we will be going over the features we are launching today.

@AutoTLDR

 

This is a game that tests your ability to predict ("forecast") how well GPT-4 will perform at various types of questions. (In caase you've been living under a rock these last few months, GPT-4 is a state-of-the-art "AI" language model that can solve all kinds of tasks.)

Many people speak very confidently about what capabilities large language models do and do not have (and sometimes even could or could never have). I get the impression that most people who make such claims don't even know what current models can do. So: put yourself to the test.

 

Increasingly powerful AI systems are being released at an increasingly rapid pace. This week saw the debut of Claude 2, likely the second most capable AI system available to the public. The week before, Open AI released Code Interpreter, the most sophisticated mode of AI yet available. The week before that, some AIs got the ability to see images.

And yet not a single AI lab seems to have provided any user documentation. Instead, the only user guides out there appear to be Twitter influencer threads. Documentation-by-rumor is a weird choice for organizations claiming to be concerned about proper use of their technologies, but here we are.

@AutoTLDR

 

TL;DR: (by GPT-4 🤖)

The article by Chandler Kilpatrick on Medium discusses the new Code Interpreter feature of ChatGPT, which has been released to Beta from its previous Alpha testing phase. The Code Interpreter enhances ChatGPT's ability to process, generate, manipulate, and run code, currently supporting only Python. Users can upload files (with a limit of 100 MB per file) for the AI to interact with, although it cannot edit files directly. The Code Interpreter can be used in various fields such as software development, data analytics, documentation, and education, helping with tasks like code generation, error detection, code refactoring, creating data visualizations, and providing real-time programming tutoring. The article also highlights some impressive feats accomplished by users, including recreating the game Flappy Bird in less than 10 minutes.

 

LLM is my command-line utility and Python library for working with large language models such as GPT-4. I just released version 0.5 with a huge new feature: you can now install plugins that add support for additional models to the tool, including models that can run on your own hardware.

@AutoTLDR

 

An AI-first notebook, grounded in your own documents, designed to help you gain insights faster.

@AutoTLDR

[–] sisyphean@programming.dev 1 points 1 year ago* (last edited 1 year ago)

TIL. Thank you! (Now I will ssh into all my VPSes and set this up!)

(cool username btw)

[–] sisyphean@programming.dev 1 points 1 year ago* (last edited 1 year ago)

This describes 99% of AI startups.

The company I work for was considering using Mendable for AI-powered documentation search. I built a prototype using OpenAI embeddings and GPT-3.5 that was just as good as their product in a day. They didn’t buy Mendable :)

[–] sisyphean@programming.dev 1 points 1 year ago

This is an excellent explanation of hashing, and the interactive animations make it very enjoyable and easy to follow.

[–] sisyphean@programming.dev 1 points 1 year ago* (last edited 1 year ago)

The problem is that they "see" the text at the token level instead of the level of characters. That's why they are bad at reversing strings or counting characters, for example. They perceive tokens as the atomic units of text instead of characters. For example, see how this comment gets tokenized:

With the token IDs shown:

The current ChatGPTs got pretty good at these tasks but they are still hard for them.

Here is an example of a (admittedly more complicated) character-level task failing:

Source: https://www.reddit.com/r/ChatGPT/comments/11z9tuk/chatgpt_vs_reversed_text/ (It's from the devil's website, so don't open it)

Related tweet by @karpathy:

https://twitter.com/karpathy/status/1657949234535211009

Text reversing example from a tweet by @npew:

EDIT: sorry for the infodump, I just find these topics fascinating.

[–] sisyphean@programming.dev 1 points 1 year ago

Is that because most of your recipes are from the US?

[–] sisyphean@programming.dev 1 points 1 year ago (2 children)

We use Celsius like for everything else

[–] sisyphean@programming.dev 0 points 1 year ago* (last edited 1 year ago) (2 children)

Well, there’s this place:

My new community got quite a few subscribers from there. Just make sure to post relative links using both the Lemmy and kbin routes (/c/ and /m/).

EDIT: oh, I almost forgot, there actually is a site for community discovery: Lemmy Browser. I don’t think it currently lists kbin communities but we could ask them to (or if it’s open source, someone could implement it).

[–] sisyphean@programming.dev 0 points 1 year ago (1 children)

These are all very useful features! Is there any chance they will get merged into the main Lemmy codebase?

[–] sisyphean@programming.dev 0 points 1 year ago (1 children)

This is pretty awesome and it shows how far .NET has come in recent years.

view more: next ›