Cheap_Meeting

Cheap_Meeting t1_j973fm1 wrote

I went ahead and asked ChatGPT your question for you:

No, Google is not a language transformer like ChatGPT. While Google has developed language models like BERT and GPT-3, it is primarily a search engine that uses various algorithms to deliver search results to users. Google's language models are used to improve search results and to power services like Google Assistant, but they are not the primary focus of the company.

As for ChatGPT, it is true that its main purpose is to generate human-like text based on prompts provided by users. However, it is not just a search engine that can talk back. It is a complex machine learning model that has been trained on vast amounts of text data and uses advanced natural language processing techniques to generate responses.

While it is true that the development of more algorithms and computing power is necessary for further advancements in AI, ChatGPT and other similar models have already made significant strides in the field of natural language processing. They have the potential to be used in a wide range of applications, including language translation, content creation, and customer service. However, it is unlikely that they will completely replace human jobs, as they are still limited by their inability to understand and reason about the world in the way that humans do.

1

Cheap_Meeting t1_j7chivx wrote

In terms of Consumer Apps, the Poe app from Quora has access to two models from Open AI and one from Anthropic.

Perplexity.ai, YouChat and Neeva are search engines that integrated LLMs.

Google has an AI + Search Event on Wednesday where they are likely to announce something as well.

In terms of APIs and getting a feeling for these models, I would use OpenAI's APIs. Their models are the best publically available models. Open Source models are still far behind.

6

Cheap_Meeting t1_j3ulqr7 wrote

Data scientist is not a very well-defined job title. It can involve more or less social interaction depending on the company.

But you could get more social interaction by doing pair programming with your coworkers or asking to be assigned more work involving social interaction, e.g. talking to clients or mentoring new coworkers.

You could eventually transition into another career path such as people management, education, project management, program management, sales, etc.

16

Cheap_Meeting t1_j21v6mi wrote

I think the main limitations of LLMs are:

  1. Hallucinations: They will make up facts.
  2. Alignment/Safety: They will sometimes give undesirable outputs.
  3. "Honesty": They cannot make reliable statements about their own knowledge and capabilities.
  4. Reliability: They can perform a lot of tasks, but often not reliably.
  5. Long-context (& lack of memory): They cannot (trivially) be used if the input size exceeds the context length.
  6. Generalization: They often require task-specific finetuning or prompting.
  7. Single modality: They cannot easily perform tasks on audio, image, video.
  8. Input/Output paradigm: It is unclear on how to use them for tasks which don't have a specific inputs and outputs (e.g. tasks which require taking many steps).
  9. Agency: LLMs don't act as agents which have their own goals.
  10. Cost: Both training and inference incur significant cost.
3

Cheap_Meeting t1_j21q96n wrote

Yes, they are trained on a much larger amount of language data than a human sees in their lifetime.

However, I would argue that it's a worthwhile trade-off. Computers can more easily ingest a large amount of data. Humans get feedback from the environment (like their parents), can cross-reference different modalities, and have inductive biases.

2

Cheap_Meeting t1_ixnt905 wrote

This reads like an announcement for the release of a traditional piece of software. It would be nice if you could instead publish some metrics such as FID or ideally side-by-side human evaluation against SD 1.5 / DALLE-2.

One of the best things about the machine learning community is that we have been taking a rational metrics-driving approach. I hope that as ML gets more and more real-world use cases, and both open-source and commercial applications that are not tied to academic research become more prevalent, we don't lose that.

17

Cheap_Meeting t1_ix56atv wrote

I disagree with some of the other advice here. I would suggest starting with something that you know works. That means you could either use a training setup from another modality such as vision or text and apply it to your data, or you could try to reproduce a result from the literature first.

2

Cheap_Meeting t1_ix520oo wrote

Rereading my own comment, it could have been phrased better. Let me try again:

I think you are taking OP's question too literally. At least as I understand it, the intent of OP's question was: "Why are self-supervised autoregressive models the predominant form of generative models for language? Intuitively it would seem that the training process should be closer to how humans learn language."

4