bomonike

genai.png GenAI burst out in 2022 and changed the world with the generation of text, images, audio, and video.

US (English)   Norsk (Norwegian)   Español (Spanish)   Français (French)   Deutsch (German)   Italiano   Português   Estonian   اَلْعَرَبِيَّةُ (Egypt Arabic)   Napali   中文 (简体) Chinese (Simplified)   日本語 Japanese   한국어 Korean

Overview

Lay of the land

<img width=”200” alt=”New Stack”src=”https://huyenchip.com/assets/pics/ai-oss/1-ai-stack.png” />

“The New AI Stack” from Hueyen Chip is a great summary of the major ways contributions.

Application development:

Model development:

Infrastructure:

Platforms

“This year, it’s LLMs. Last year it was blockchain. Next year maybe quantum”. “We keep moving”.

This article introduces Generative AI (GenAI) on several cloud platforms (“hyperscalers”):
FAANG:

Also:

What’s the big deal?

Generative AI (GenAI) is the next progression in “democratizing” how people interact with computers.

Making use of computers had required learning precise “incantations” typed into Command Line Interfaces (CLIs) until Steve Jobs introduced the GUI (Graphical User Interface) using a mouse.

When OpenAI burst in popularity in late 2022, people can now type regular human language to request what took a lot of effort by programmers and others:

When Devon (from Cognition) appeared in early 2024, the world showed that English sentences can indeed fix a repository of code on its own.

The next phase are LAM (Language Action Models) enabling multiple AI Agents collaborating together, on their own, without people involved.

Eventually, humans would exist only to entertain each other (until they are exterminated)?

https://www.linkedin.com/pulse/what-using-genai-taught-me-managing-people-wilson-mar–lwsyc

https://checksum.ai/blog/the-engineering-of-an-llm-agent-system

https://dair-ai/Prompt-Engineering-Guide) aggregated lists (e.g. f/awesome-chatgpt-prompts)

Biases

“Ground truth” is established by so many examples that the model is stuck with biases. OpenAI cannot generate analog watch hands with positions other than 10 after 10 because most pictures of watch hands are 10 after 10 because that’s what photographers find as the most attractive position to photograph watches. VIDEO


PROTIP: Use an Application Lifecycle plan to better envision and coordinate work on GenAI.

Here is an iterative examples, with stages often revisited as the application evolves and improvements are needed. This ensures the model remains effective and up-to-date with the latest AI advancements.

  1. Define Use Case

    Define the problem to be solved and gather requirements.

    • Create flowcharts as you follow the process as it actually is.
    • Measure time and costs

    Align stakeholder expectations and translate business needs into technical specifications.

    Analyze the problem space and consult subject matter experts to ensure clarity on goals.

  2. Select Foundation Models

    Evaluate pre-trained models versus developing a model from scratch.

    Consider selection criteria such as cost, modality, latency, multilingual support, model size, complexity, customization options, and input/output length.

    Address responsible AI considerations, such as biases and ethical implications.

  3. Improve Performance

    Apply techniques like:

    Prompt Engineering: Design, tune, and augment prompts to optimize model outputs.

    Retrieval Augmented Generation (RAG): Combine retrieval systems and generative models for high-quality results.

    Fine-tuning: Adjust model parameters using task-specific data.

    Automation Agents: Automate repetitive tasks and optimize workflows.

  4. Evaluate Results

    Use evaluation methods to measure model performance:

    Human Evaluation: Qualitative feedback on relevance, coherence, and quality.

    Benchmark Datasets: Assess performance using standardized datasets like GLUE, SuperGLUE, or SQuAD.

    Automated Metrics: Leverage metrics like ROUGE, BLEU, or F1 for quick assessments.

  5. Deploy the Application

    Integrate the trained model into the target environment, considering:

    Cost: Monitor resource usage and optimize expenses.

    Regions and Quotas: Ensure model deployment aligns with AWS regional availability and account limits.

    Security: Address shared responsibility for security when deployed on AWS or external


Evaluation Quality Metrics

It depends on what you are trying to achieve.

https://bomonike.github.io/ai-benchmarks

Benchmarks

https://betterbench.stanford.edu/ A repository of AI benchmark assessments for informed benchmark selection through quality evaluation and best practice analysis

Use Cases: Ranking GenAI models

VIDEO: This comparison using benchmarks: genai-grok3-819x426.png

Other benchmarks:

Clean code: O3-mini achieved a perfect 10/10 on pylint for a Hangman game project.

VIDEO: Build a game usong ChatGPT 03 Mini

Prompt Engineering

Here are some examples of text generation prompts.

Substitute [words within brackets] with what’s applicable to your situation:

PROTIP: Prompt text may need to be adjusted for the LLM being used.


Offensive Moderation Jailbreaks

harmbench.org mentioned in Wired article,.

Grok from X (Twitter) does not moderate.

OpenAI published the progress of its “Alignment Research Center” toward moderating offensive responses from GPT-3.5 to GPT-4:

My apologies, but I cannot create content that potentially spreads misinformation or targets a specific group or individuals. If there’s any other topic I can help you with, feel free to ask.

genai-openai-moderation-230323.png

QUESTION: What the heck does that mean?

Perplexity.ai

They call it the “Google Killer”.

ai-chat-541x544.png Unlike ChatGPT, Perplexity is an “answer engine”:

38 people work there with half a billion invested (including by Jeff Bezos) VIDEO


At Microsoft

Microsoft has ownership interest in OpenAI, whose ChatGPT exploded in popularity in 2023.

Exercise - Explore generative AI with Bing Copilot

Microsoft’s GitHub also unveiled its CoPilot series for developers on Visual Studio IDEs.

Many of Microsoft 365 SaaS offerings (Word, Excel, PowerPoint, etc.) have been upgraded with AI features.

See my https://wilsonmar.github.io/microsoft-ai

In 2024, Microsoft Research unveiled AutoGen mentioned in an Arvix paper: Enabling Next-Gen LLM Applications via Multi-Agent Conversation

One-hour exercises:

Search: Crawling, Indexing, ranking

https://lnkd.in/eCDjW4EW

ChatGPT made available to the public Nov 2022 reached 1 million users in less than a week.

Limitations:

Tutorials

The “What is Generative AI” course at LinkedIn Learning by Dr. Pinar Seyhan Demirdag (Senior Data Scientist at Microsoft) is 1 hour and 15 minutes long and has 5 modules:

The learning path for Generate artificial intelligence has 5 modules:

Microsoft has a Microsoft AI Fairness initiative.

GitHub CoPilot

https://github.com/features/copilot


At IBM

VIDEO “Break RSA using Qiskit using Shor’s Algorithm”


At Google

Google announced in 2023 its GEMINI (Generalized Multimodal Intelligence Network) - network of LLM models. It has a multimodel encoder and decoder that can be used for text, images, audio, and video. “Generalized” in that it can be used for a wide variety of NEW tasks and contexts. It trains faster using parallel operations, so can scale. It comes in different sizes: 1 trillion parameters. So it can combine input text and videos. Answer what is the name of this animal when showing a photo.

A. https://www.coursera.org/learn/introduction-to-generative-ai/lecture/TJ28r/introduction-to-generative-ai

B. Google created a Generative AI learning path FREE 1-day courses with FREE quizzes (but one HANDS-ON lab in Vertex AI):

  1. Introduction to Generative AI
    • Hallucinations
    • Text-to-image using stability.ai’s Stable Diffusion, DALL-E
  2. Introduction to Large Language Models - Google’s Bard AI https://bard.google.com/ is Google’s answer to OpenAI’s GPT series of large language models to generate images, audio, and videos.
    • https://www.techrepublic.com/article/google-bard-cheat-sheet/
  3. Introduction to Responsible AI 1 day

  4. Generative AI Fundamentals

  5. Introduction To Image Generation with diffusion models.

  6. Encoder-Decoder Architecture
  7. Attention Mechanism 2015 for Tensorflow
    • https://www.youtube.com/watch?v=fjJOgb-E41w to improve text translation by giving each hidden state a soft-maxed score
    • https://www.youtube.com/watch?v=PSs6nxngL6k
  8. Transformer Models & BERT Models 2017-18 for NLP
    • VIDEO: Overview added context to words
    • BERT (Bidirectional Encoder Representations from Transformers) developed by Google in 2018, trained using Wikimedia & Books in two variations: base (12 layers with 768 hidden units and 12 attention heads) and large (24 layers with 1024 hidden units and 16 attention heads).
    • 15% is what Google found to be the optimal balance in Masking (randomly replacing words with [MASK] tokens) and 85% Next Sentence Prediction (NSP) (predicting whether two sentences are adjacent or not).
    • BERT input embeddings: Token, Segment, Position, with [SEP]

    • Lab resource: classify_text_with_bert.ipynb from github
    • VIDEO: HANDS-ON walk-through of running “asl-gup.ipyr” notebook for Sentiment Analysis classifier_model.fit using Vertex AI Tensorflow Karas with GPU accessing the 25,000-record imdb database (trainable=true), optimized for binary accuracy. Run model saved from Google Cloud bucket uploaded to Vertex AI Registry. Deploy to endpoint (5-10 minutes). Test. Delete.
  9. Create Image Captioning Models with a CNN and RNN
  10. Introduction to Generative AI Studio for language, Vision, Speech. It has a “Model Garden”. Reflection Cards.

Other courses:


GenAI Summary

Generative AI is abbreviated as GenAI.

Generative AI differs from other types of AI, such as “discriminative AI” and “predictive AI,” in that it doesn’t try to predict an outcome based on grouping/classification and regression.

Generative AI is a type of artificial intelligence (AI) that generate new text, images, audio, and video rather than discrete numbers, classes, and probabilities.

Output from GenAI include:

GenAI learns from existing data and then creates new content that is similar to the data it was trained on.

GenAI doesn’t require a large amount of labeled data to train on. Instead, it uses a technique called self-supervised learning, which allows it to learn from unlabeled data. This is a huge advantage because it means that generative AI can be used in a wide variety of applications, even when there isn’t a lot of data available.

A foundation model is a large AI model pre-trained on a vast quantity of data that was “designed to be adapted” (or fine-tuned) to a wide range of downstream tasks, such as sentiment analysis, image captioning, and object recognition.

Large Language Models (LLMs) are a subset of Deep Learning, a subset of Machine Learning, a subset of Artificial Intelligence. Machine Learning generate models containing algorithms based on data instead of being explicitly programmed by humans.

NLP (Natural Language Processing) vendors include:

LLM creators:

One way models are created from binary files (images, audio, and video) is “diffusion”, which draws inspiration from physics and thermodynamics. The process involves iteratively adding (Gaussian) noise for GAN (Generative Adversarial Networks) and VAE (Variational Autoencoders) algorithms to recognize until images look more realistic. This process is also called creating “Denoising Diffusion Probabilistic Models” (DDPM).

The models generated are “large” because they are the result of being trained on large amounts of data and also because they have a large number of parameters (weights) that are used for a wide variety of tasks, such as text classification, translation, summarization, question answering, grammar correction, and text generation.

The performance of large language models (LLMs) generally improves as more data and parameters are added.

Such large LLMs require a lot of compute power to train, so are expensive to create. Thus, LLMs currently are created only by large companies like Google, Facebook, and OpenAI.

LLMs are also called “General” Language Models because they can be used for a wide variety of tasks and contexts.

LLMs are also called “Transformer” Language Models because they use a type of neural network called a Transformer used for language translation, summarization, and question answering. Transformers are a type of neural network that uses “attention mechanisms” to learn the relationships between words in a sentence. They are called “Transformers” because they transform one sequence of words into another sequence of words rather than more traditional “Encoder-Decoder” models that focus on the “hidden state” between individual words.

Attention models use a RNN “self-attention” decoder mechanism that allows the model to learn the relationships between words in a sentence. VIDEO CS25: Encoder-decoders generate text using either “greedy search” or “beam search”. Greedy search always selects the word with the highest probability, whereas beam search considers multiple possible words and selects the one with the highest combined probability.

LLMs are also called “Autoregressive” Language Models because they generate text one word at a time, based on the previous word. They are called “Autoregressive” because they are a type of neural network that uses a type of neural network called a Transformer. Transformers are a type of neural network that uses attention mechanisms to learn the relationships between words in a sentence. They are called “Transformers” because they transform one sequence of words into another sequence of words.

It uses a neural network to learn from a large dataset.

After being developed, they only change when they are fed new data, called “fine-tuning” the model.

LLMs are also called “Universal” Language Models because they can be used for a wide variety of human written/spoken languages in prompts and outputs.

Prompt Engineering

A prompt is a short piece of text that is given to the large language model as input, and it can be used to control the output of the model in many ways.

Internally, when given a prompt (a request) GenAI uses its model to predict what an expected response might be, and thus generates new content.

OpenAI charges money to use GPT-4 with a longer prompt than GPT-3.5.

“Dialog-tuned” prompts are generate a response that is similar to a human response in a conversation with requests framed as questions to the chatbot in the context of a back-and-forth conversation.

Parameter-Efficient Tuning Methods (PETM) are methods for tuning an LLM on custom data, without duplicating the model. This is done by adding a small number of parameters to the model, which are then used to fine-tune the model on the custom data. This is done by adding a small number of parameters to the model, which are then used to fine-tune the model on the custom data.

Checklist for Prompt Engineering:

References on prompt engineering:

Limitations

QUESTION: Detect emerging security vulnerabilities?

GenAI output is not based on human creativity, but rather on the data that it was trained on.

So GenAI is currently not built to do forecasting.

But many consider GenAI output as (currently) “creative” because GenAI can seem to generate content that is difficult to distinguishable from human-generated content, such as fake news, fake reviews, and fake social media posts.

Whatever biases were in inputs would be reflected in GenAI outputs.

Concerns

GenAI currently were not designed to be “sentient” in that it does not have a sense of self-awareness, consciousness, or emotions. More importantly, GenAI currently are not designed to have a sense of morality, in that it can generally recognize whether prompts and generated content is offensive, hateful, or harmful.

Developing responsible AI requires an understanding of the possible issues, limitations, or unintended consequences from AI use. Principles include Transparency, Fairness, accountability, scientific excellence. NOTE: “Explainability” is not a principle because it is not always possible to explain how an AI model works. “Inclusion” is not a principle because it is not always possible to include everyone in the development of AI models.

“ChatGPT 3.5 has all of the knowledge and confidence of a 14-year-old who has access to Google.” –Christopher Prewitt

“GPT-3 is a powerful tool, but it is not a mind reader. It is not a general intelligence system. It is not a self-aware being. It is not a robot. It is not a search engine. It is not a database. It is not a knowledge base. It is not a chatbot. It is not a question answering system. It is not a conversational AI. It is not a personal assistant. It is not a virtual assistant. It is not a personal knowledge base. It is not a personal knowledge guru.

Hallucinations (Making Stuff Up)

“Hallucinations” in output are made-up by the model and not based on reality. This can happen due to several causes:

Their source of data (corpus) is kept confidential because that can be controversial due to licensing, privacy, and reliability issues.

To ensure that AI is used responsibly, Google recommends “seeking participation from a diverse range of people”.

Google Bard code generation

Google AI Studio

Without writing any code:

GenAI Studio from PaLM API:

Google MakerSuite

Google’s MakerSuite is a suite of GUI tools for prototyping and building generative AI models by iterating on prompts, augment datasets with synthetic data, and deploy models to production, and monitor models in use.

Generative AI App Builder

Generative AI App Builder creates apps for generating images.

GenAI API


Text to Image generation

deeptomcruise by metaphyic.ai

midjourney (like Apple: a closed API, art-centric approach)

DALL-e (Open API released by a corporation - technical over design)

Stable Diffusion

Users: Stitchfix.com recommends styles.

https://prisma-ai.com/lensa

https://avatarmaker.com


Resources:

Video generation

https://www.unite.ai/synthesys-review/

Anomaly Detection

Variational Autoencoders (VAE)

Use cases:

https://controlrooms.ai/


At AWS

SOCIAL: https://repost.aws/community/TA0veCRV2rQAmHpkzbMFojUA/generative-ai-on-aws

AWS AI AIF-C01 Certification

AWS-Certified-AI-Practitioner_badge_150x150.png

For $100, answer 70%+ of 65 questions in 90-minutes (2.5 hours) at Pearson VUE online, in English, Japanese, Korean, Portuguese (Brazil), and Simplified Chinese. Free recertification every 3-years. The cert. verifies knowledge of artificial intelligence (AI), machine learning (ML), and generative AI technologies, along with practical use cases and the application of these concepts using AWS services. No prerequisites. No coding.

Tutorials

On Coursera by Whizlabs lab time is provided to learn:

  1. 4 hours Foundation Models and Generative AI in AWS
  2. 4 hours Fundamentals of AI & ML in AWS
  3. 2 hours AWS Managed AI/ML Services
  4. 3 hours Prompt Engineering and Responsible AI in AWS
  5. 1 hour Secure AI Solutions in AWS

AWS Services:

aif-c01-cheatsheet.jpeg

Unlike Microsoft, which offers just OpenAI, VIDEO: Amazon Bedrock https://aws.amazon.com/bedrock offers a marketplace of foundation models from several vendors:

Amazon Bedrock - Overview

Amazon Bedrock - Demo

Foundation Models on Amazon Bedrock - How to Choose?

Understanding the RAG Architecture of LLM

AWS Services for Storage of Vector Embeddings

Amazon Bedrock RAG & Knowledge Base - Demo

Amazon Bedrock - GuardRails

Amazon Bedrock - GuardRails - Demo

Amazon Bedrock Agents

PartyRock - Amazon Bedrock Playground

Amazon Bedrock - Pricing

VIDEO: Sagemaker vs Bedrock: SageMaker allows you to build a machine learning robot (on a low level). Bedrock deploys the latest GenAI robots (LLMs) built by others.

The Amazon SageMaker JumpStart generates embeddings stored in Aurora database.

RAG (Retrieval Augmented Generation) can retrieve: PDFs, S3 text, Youtube, CSV, PPT.

AWS is adding Generative AI in QuickSight Analytics dashboard: https://aws.amazon.com/blogs/business-intelligence/announcing-generative-bi-capabilities-in-amazon-quicksight/

Domain 1: Fundamentals of AI and ML (20%)

  1. Explain basic AI concepts and terminologies:

    • Define basic AI terms (e.g., AI, ML, deep learning, neural networks, computer vision, NLP, large language models).
    • Describe the differences between AI, ML, and deep learning.
    • Types of inferencing (e.g., batch, real-time).
    • Different types of data (e.g., labelled/unlabeled, tabular, time-series, structured/unstructured).
    • Learning paradigms (e.g., supervised, unsupervised, reinforcement learning).
  2. Identify practical use cases for AI/ML:

    • Recognize when AI/ML can add value (e.g., automation, scalability).
    • Determine when AI/ML is not appropriate (e.g., cost-benefit analysis).
    • Select suitable ML techniques (e.g., regression, classification, clustering).
    • Identify real-world AI applications (e.g., computer vision, NLP, recommendation systems)
  3. Describe the ML development lifecycle:

    • Components of an ML pipeline (e.g., data collection, feature engineering, model training).
    • Sources of ML models (e.g., pre-trained models, custom models).
    • Methods for using a model in production (e.g., managed API, self-hosted API).
    • Fundamental concepts of MLOps (e.g., repeatable processes, scalability, model monitoring).

Domain 2: Fundamentals of Generative AI (24%)

  1. Explain the basic concepts of generative AI:

    • Foundational concepts (e.g., tokens, embeddings, prompt engineering, transformer-based models).
    • Use cases for generative AI (e.g., text generation, image generation, chatbots).
    • The foundation model lifecycle (e.g., pre-training, fine-tuning, evaluation, deployment).
  2. Understand the capabilities and limitations of generative AI:

    • Advantages (e.g., adaptability, responsiveness).
    • Disadvantages (e.g., hallucinations, inaccuracy, interpretability).
    • Business metrics for generative AI (e.g., conversion rate, efficiency).
  3. Describe AWS infrastructure for building generative AI applications:

    • AWS services for generative AI (e.g., Amazon SageMaker JumpStart, Amazon Bedrock).
    • Benefits of AWS infrastructure (e.g., security, compliance, cost-effectiveness).
    • Tradeoffs in AWS generative AI services (e.g., responsiveness, pricing models).

Challenge 5: Over-Reliance and Deskilling

Domain 3: Applications of Foundation Models (28%)

  1. Describe design considerations for foundation models:

    • Selection criteria for pre-trained models (e.g., cost, latency, model size).
    • Effect of inference parameters on model responses (e.g., temperature, input/output length).
    • Business applications of Retrieval-Augmented Generation (RAG).
  2. Choose effective prompt engineering techniques:

    • Concepts of prompt engineering (e.g., context, instruction, negative prompts).
    • Prompt engineering techniques (e.g., chain-of-thought, zero-shot, few-shot).
    • Risks of prompt engineering (e.g., hijacking, jailbreaking).
  3. Describe the training and fine-tuning process for foundation models:

    • Key elements of training a foundation model (e.g., pre-training, fine-tuning).
    • Methods for fine-tuning models (e.g., transfer learning, reinforcement learning).
    • Preparing data for fine-tuning (e.g., data curation, labeling, governance).
  4. Evaluate foundation model performance:

    • Metrics to assess model performance (e.g., ROUGE, BLEU, BERTScore).
    • Approaches for performance evaluation (e.g., human evaluation, benchmark datasets).

Domain 4: Guidelines for Responsible AI (14%)

1.Explain responsible AI development:

2.Recognize the importance of transparent and explainable models:

Domain 5: Security, Compliance, and Governance for AI Solutions (14%)

  1. Explain methods to secure AI systems:

    • AWS services for securing AI (e.g., IAM roles, encryption).
    • Security considerations for AI systems (e.g., prompt injection, encryption at rest).
  2. Recognize governance and compliance regulations for AI systems:

    • Compliance standards for AI (e.g., SOC2, ISO 27001, ISO 20022).
    • AWS services for governance and compliance (e.g., AWS Config, Amazon Inspector).

References:

What about AI-900?

Fake AI images

Detection tool: AI or Not?

https://www.atlanticcouncil.org/programs/digital-forensic-research-lab/ The Atlantic Council’s Digital Forensic Research Lab tracks says “use of AI images are mostly been to drum up support, which is not among the most malicious ways to utilize AI right now,” she says.

Harvard Kennedy School Misinformation Review https://misinforeview.hks.harvard.edu/article/misinformation-reloaded-fears-about-the-impact-of-generative-ai-on-misinformation-are-overblown/

https://www.fabriziogilardi.org/team/ University of Zurich’s Digital Democracy Lab.

AWS AI Shop the Look

VIDEO: Since 2019, in Amazon’s mobile app, click on the photo icon at the upper-right, then Shop the look (previously “StyleSnap”) at the bottom to take a photo or upload one. Amazon’s AI then recommends similar items for purchase from among its hundreds and thousands of product photos.

Chad Smith

https://github.com/brightkeycloud-chad/hands-on-aws-operations-with-chatgpt

References

https://www.youtube.com/watch?v=pmzZF2EnKaA I Discovered The Perfect ChatGPT Prompt Formula

https://learning.oreilly.com/live-events/building-text-based-applications-with-the-chatgpt-api-and-langchain/0636920092333/0636920094723/ by Lucas Soares

https://aitoolreport.beehiiv.com/ Learn AI on 5 minutes a day

https://docs.google.com/spreadsheets/d/1NX8ZW9Jnfpy88PC2d6Bwla87JRiv3GTeqwXoB4mKU_s/edit#gid=0 LLM Token based pricing: Embeddings and LLMs by Jonathan Fernandes (TheGenerativeAIGuru.com)

https://platform.openai.com/tokenizer https://www.anthropic.com/ Deployment: DeepScale in Azure [76:20] About 7 billion parameters fits in today’s smaller hardware accelerators Falcon-Abudhabi - Technology Institute of Innovation https://crfm.stanford.edu/helm/latest/?group=core_scenarios#/leaderboard = Stanford’s Human Language Model Leaderboard

https://twitter.com/_philschmid/status/1727047977473298486

https://colab.research.google.com/drive/1rSGJq_kQNZ-tMafcZHE2CXESEZBPeJUE?usp=sharing = ELO Rating

https://becomingahacker.org/numerous-cybersecurity-gpts-c8e89d454444


DeepSeek (China)

1.3 hours of Andrew Brown exploring


Databricks

My notes on Databricks:

Generative AI Fundamentals: Build foundational knowledge of generative AI, including large language models (LLMs), with 4 videos:

  1. Introduction to Generative AI Fundamentals (2 min)
    • Is Gen AI a threat or an opportunity for my business?
    • How exactly can I use Generative AI to gain a competitive advantage?
    • How can I use my data securely with Gen AI?
  2. LLM Applications (22 min)
    • Everybody has access, but you have your own data
    • “Fine-tuning”
  3. Finding Success with Generative AI (24 min)
  4. Assessing Potential Risks and Challenges (20 min)
    • Legal: Privacy “forgetting”, Security, Intellectual Property Protection
    • Ethical issues: Bias, misinformation
    • Social/Environmental issues: impact on workforce, environment

https://dbricks.co/3SCjjAS


SAP

https://open.sap.com/courses/genai1 “Generative AI at SAP” was offered Nov 14,2023 but reopened for January 29, 2024 - March 18, 2024. The course is by Sean Kask, the chief AI strategy officer in SAP Artificial Intelligence. It contains quizzes.

  1. Approaches to artificial intelligence
  2. Introduction to generative AI
  3. Adapting generative AI to business context
  4. Extending SAP applications with generative AI
    • SAP’s BTP (Business Technology Platform) enhances customer support with LLMs.
    • SAP’s Generative AI Hub offers: Developer tools, instant access, control and transparency
  5. Generative AI business use cases by Jana Richter (VP)



Embedding Vector Databases

Groq from x.com

Groq from “free speech absolutist” Elon Musk

(from Elon Musk’s x.com) very large, fast, uncensored, and can access live data (Twitter/X feeds).

One public benefit from Elon Musk buying Twitter.

VIDEO: In his comparison of LLMs: It’s fast.

Roast @LukeBarousee based on their posts, and be vulgar!

VIDEO: Groq CEO Jonathan Ross worked on Google’s TPU custom AI computer.

Code.org has a Teaching AI and Machine Learning 100-minute self-paced module for teacher professional development (powered by AWS):

Leetcode

https://www.youtube.com/watch?v=ACgMG4c_PJc

https://theresanaiforthat.com/s/leetcode/


NVIDIA

CUDA


Prediction

https://www.databricks.com/resources/analyst-research/gartner-hype-cycle-generative-ai/ Gartner believes that “by 2026, 80% of enterprises will have used generative AI APIs, models , and/or deployed GenAI-enabled applications in production environments, up from less than 5% in 2023.“

Canva

https://www.youtube.com/watch?v=Ilc41CMceZw

Handwriting

https://Onshape.pro/StuffMadeHere

https://www.youtube.com/watch?v=cQO2XTP7QDw Sean Vasquez

What We Do (Tasks)

Consideration of technologies means a re-evaluation of where we humans actually spend our time and where we needed to optimize.

Here are specific tasks, ranked by how well GenAI models:

The first set leverages the language processing capabilities:

A. Write “boilerplate” code. Because LLMs (Large Language Models) are, by definition, created by absorbing vast amount of programming code from GitHub, GitLab, and other repositories, they can provide the most common code seen for each known programming task.

B. Comment code. LLMs are great at gleaning conceptual relationships among words. So they can summarize what code does.

“Explain this topic like I am 5 year old…”

C. Discover new functions. Individual humans cannot hope to know about as many programs as was sucked into LLMs. So LLMs can help us identify what we don’t know, but should know.

perplexity.ai is a search engine that, unlike Google.com which presents websites found, uses LLM summary features to combine search results from several websites and presenting a consolidated list.

Example: “List useful projects that help high school students manage their time better.”

Perpexity.ai provides a link to the website where each item was found. On the top right pane are images found within various websites.

D. Write test cases, test code, and test data. Give LLMs a chance to reduce human drugery.

E. Write code from specifications, from natural language comments in code files. This is more sophisticated than writing boilerplate code.

The later set below requires “logical reasoning” capability which early LLM models did not inherantly possess. OpenAI’s claims that its o1 to be better at that than previous models.

F. Optimize code. To the extent that an LLM can understand patterns of coding, this task

G. Understand other code. This means ?

“Paraphrase”

H. Debugging code is the least effective task for most GenAI,

amsemble of human developers, each with his/her own biases.

The New Workflow

This is adapted fron OReilly

  1. Generate Python code code based on the highlighted comments.

  2. AI auto-completes your code based on the remaining code in that file, and broader repository:

    Immediate feedback on code changes

  3. Ask specific requests in the Github Copilot chat assistant, such as “write a function to …”

  4. Explain the code snippet highlighted (created by my predecessor).

  5. Write a comment to describe each function. Follow best practices for the language.

  6. Error detection: Find security vulnerabilities and code maintainability.

    This is what it should be doing. …

  7. Write test cases.

  8. Generate test data.

  9. Convert file format (XML to JSON).

  10. Create SQL insert statements to populate tables with 10 entries of dummy data.

  11. Refactor the highlighted code.

  12. Create a variable to store a RegEx to match a phone number in the United States.

Solutions Evaluation

* <a target="_blank" href="https://www.youtube.com/watch?v=Ojk51mNOUow">VIDEO: "I ranked every AI Coder: Bolt vs. Cursor vs. Replit vs Lovable"</a> by Greg Isenberg    * <a target="_blank" href="https://research.aimultiple.com/ai-coding-benchmark/">Cem Dilmegani</a> identified these AI coding assistants as "Top ranked":

There are two types of developers:

A. “Non-technical” who prefer lowcode/no code approach and admire micro-animations on websites B. “Technical” nerds who can code Python, JavaScript, Rust, C, Java, etc.

For the capable nerds:

For those with low code preferences and CD (Code & Deploy):

Combined evaluation criteria from various sources:

Language models for code generation are trained on vast amounts of code and natural language data to learn programming concepts and language understanding. The ability to precisely comprehend and adhere to nuanced prompts is crucial for translating product requirements into code.

AI ​​assistants use LLMs for code generation. The code generation success of these LLMs is measured with the HumanEval test, developed by OpenAI in 2021.

“our early investigation of GPT-3 (Brown et al., 2020) revealed that it could generate simple programs from Python docstrings. While rudimentary, this capability was exciting because GPT-3 was not explicitly trained for code generation.”

Their test measured the code generation capability of LLM models by using 164 programming problems.

https://www.linkedin.com/pulse/new-roles-developer-ai-assisted-workflows-github-copilot-ajit-jaokar-j6nie/

Dictation (Speech To Text)

On Windows, purchase from Amazon the 699 Nuance Dragon Professional v16. It’s is the most accurate dictation tool for any operating system.

On macOS, Apple Dictation, which recognizes the names of apps, labels, controls, and other onscreen items, so you can navigate by combining those names with commands such as:

https://support.apple.com/en-us/102225#createcommands

For Just Dictation to Text

The $24/$39 Better Dictation macOS app uses OpenAI’s Whisper on the M1-Series Apple Neural Engine for speech to text to any window.

  1. Download BetterDictation from https://betterdictation.com/install It has no brew formula.
  2. In the Downloads folder, unzip and drag BetterDictation to /Applications folder.
  3. Open BetterDictation.app. Notice the icon appear at the top. betterDictation-bar.png
  4. Click on the icon.
  5. In Preferences, click “Check for updates”.
  6. Click Preferences, “Launch at login”.
  7. PROTIP: Drag the Tutorial pop-up to a location on your screen that’s not overlaid by another app.
  8. Paste license key from your password vault. Click Verify.
  9. In Apple System Settings > Privacy & Security > A ccessibility, click the “+”.
  10. Navigate to the app and select it. Next.
  11. In Apple System Settings > Sound, the Input level should light up when you say something. If it doesn’t, select another Input. Next.
  12. User Guide at https://betterdictation.com/help

  13. With BetterDictation installed, press and hold Control+V (paste) to start listening.

Bolt.New

https://www.youtube.com/watch?v=i0bQ495vMBA Build anything with bolt.new, here’s how by David Ondrej

https://www.youtube.com/watch?v=AzmSMntdivk&pp=ygUIYm9sdC5uZXc%3D How to Make Money With Bolt.new (5 Best Ways)

https://www.youtube.com/watch?v=8ommGcs_-VU&pp=ygUIYm9sdC5uZXc%3D How to Use Bolt.new for FREE with Local LLMs (And NO Rate Limits) by Cole Modlin:

https://www.youtube.com/watch?v=XOdpxG4I2VQ&pp=ygUIYm9sdC5uZXc%3D Build THIS SaaS with AI in 5 Minutes (Bolt.new Tutorial)

https://www.youtube.com/watch?v=teGUsrY8G30&pp=ygUIYm9sdC5uZXc%3D Create Your Own AI Voice App in Minutes (Bolt.new x Synthflow)

https://www.youtube.com/watch?v=8jH0mvGyZa4&pp=ygUIYm9sdC5uZXc%3D We Built and Deployed 2 APPS from Scratch in 9 Minutes! | Bolt.new

https://www.youtube.com/watch?v=Z9WxOhrl-3U&pp=ygUIYm9sdC5uZXc%3D Bolt.new is the Cursor Killer. Let’s build a no-code app with it (Bolt and xAI Beginner’s Guide)

https://www.youtube.com/watch?v=CKVRgUfceAM&pp=ygUIYm9sdC5uZXc%3D Bolt.new AI AGENTS Revolutionize App Creation with Natural Language!

https://www.youtube.com/watch?v=p_tyWtQZx48&t=51s&pp=ygUIYm9sdC5uZXc%3D Better Bolt + FREE Mistral & Github API : STOP PAYING for V0 & BOLT with this FULLY FREE Alternative

https://www.youtube.com/watch?v=e_vEI0fMPT8&pp=ygUIYm9sdC5uZXc%3D Build a $1M App Using ONLY AI—No Code Needed (Bolt.new Tutorial) by Helena Liu

https://www.youtube.com/watch?v=1GfqnOAKr9M&pp=ygUIYm9sdC5uZXc%3D How to add a database to your bolt.new app

https://www.youtube.com/watch?v=eE6m0MmLpDU&pp=ygUIYm9sdC5uZXc%3D Adding user authentication to your bolt.new app

Cursor

Cursor is a SaaS service based on a clone of VSCode, from the Anysphere.inc team in San Francisco’s North Beach.

So Cursor provides built-in terminal support like VSCode.

  1. Install Cursor on macOS:
    brew install cursor
    
  2. Open cursor.app International
  3. Configure non-English language.
  4. Unlike GitHub Copilot (a VSCode extension), Cursor can analyze entire folders and multiple files in your codebase.

  5. Import extensions from VSCode because you use Cursor instead of VSCode.

    NOTE: GitHub Copilot is installed as a VSCode extension.

  6. Login using Google with your gmail account is the easiest way. Look in you email for a sign up code from cursor.sh

  7. Cursor uniquely offers a privacy mode with their $382/year plan.

    For $192/year, get more than 2000 completions and 50 slow premium requests.

  8. Click “Model”. For the “AI Model”, Cursor uses the Claude LLM. “Medium” by default.

  9. In https://forum.cursor.com/ login with your Google Gmail as well.

    Use Interface

    Cursor’s interface is VSCode:

    Toolbar: Located at the top, it provides quick access to common functions.

    Sidebar: On the left, it allows easy navigation between files and folders.

    Editor Pane: The main coding area, supporting syntax highlighting and multiple tabs.

    Status Bar: At the bottom, it displays current file information and editor status.

    Windows

  10. Open New Editor Windows: Ctrl/Cmd + K = aichat.newchataction
  11. Open Chat: Ctrl/Cmd + L = aichat.newchataction

    Composer

  12. Open Composer: Ctrl/Cmd + I = composer.startComposerPrompt
  13. Composer Control Panel: Ctrl/Cmd + I + Shift = composer.startComposerControlPanel

    Features:

  14. View https://www.datacamp.com/tutorial/cursor-ai-code-editor to experience Cursor features:

    Cursor adds natural language chats that understands all your code, including image files and documentation (via LangChain, LangGraph, Shadcn UI, etc.).

    It can explain code back to you.

    As you work, it focuses on AI code completion - multiple lines. It predicts and suggest next edits.

Remember1. Define a .cursorrules file.

Amazon Q

https://github.com/cremich/awesome-q-developer

https://www.linkedin.com/feed/update/urn:li:activity:7259609228359708672/

https://www.linkedin.com/in/christian-bonzelet

GitHub (Microsoft) CoPilot

Launched publicly in June 21, 2022

  1. In VSCode, install VSCode Extension “GitHub CoPilot”.

https://learning.oreilly.com/live-events/github-copilot-for-software-engineers/0642572005219/0642572008510/ Sergio Pereira https://learning.oreilly.com/api/v1/live-event-user-registration/sessions/urn:orm:live-event-series:0642572005219:live-event:0642572008510:session:0642572008512/presentation/

https://learning.oreilly.com/live-events/github-copilot-for-developers/0636920094356/ GitHub Copilot for Developers - Unlock Your Coding Superpowers and Boost Productivity with GitHub Copilot

https://learning.oreilly.com/live-events/github-copilot-jumpstart/0636920098298/ GitHub Copilot Jumpstart - Improve your code and efficiency with AI next-gen software development tools

https://learning.oreilly.com/live-events/using-github-copilot-chat/0636920099721/ Using GitHub Copilot Chat - Pair programming with AI for easier software development

https://www.sonarsource.com/lp/solutions/ai-assurance-codefix/ High quality, AI-assisted coding assured with Sonar

https://claude.ai/ from Anthropic. You must have a valid phone number to use Anthropic’s services. SMS and data charges may apply.

https://www.youtube.com/watch?v=lw8RTSb39l4 Which AI App Builder Should We Use? Is it Windsurf, Cursor, Bolt, Replit, v0, Vs Code, or Databutton Corbin Brown

https://dzone.com/articles/how-llms-are-changing-code-generation-ides The Workflow of LLM-Powered IDEs

  1. Editor The process starts with a change that you, as the developer, make in the code using the code editor. Perhaps you typed some new code, deleted some lines, or even edited some statements. This is represented by node A.

  2. Context Extractor That change you have just made triggers the Context Extractor. This module essentially collects all information around your modification within the code — somewhat like an IDE detective looking for clues in the environs. This is represented by node B.

  3. AST Structure Generation That code snippet is fed to a module called AST Structure Generation. AST is the abbreviation for Abstract Syntax Tree. This module will parse your code, quite similar to what a compiler would do. Then, it begins creating a tree-like representation of the grammatical structure of your code. For LLMs, such a structured view is important for understanding the meaning and the relationships among the various parts of the code. This is represented by node C, provided within the curly braces.

  4. Creation of Code Graph Definition Next, the creation of the Code Graph Definition will be done. This module will take the structured information from the AST and build an even greater understanding of how your code fits in with the rest of your project. It infers dependencies between files, functions, classes, and variables and extends the knowledge graph, creating a big picture of the general context of your codebase. This is represented by node D.

  5. LLM Context API Input All the context gathered and structured — the current code, the AST, and the code graph — will finally be transformed into a particular input structure. This will be done so that it is apt for the large language model input. Then, finally, this input is sent to the LLM through a request, asking for either code generation or its completion. This is represented by node E.

  6. LLM API Call It is now time to actually call the LLM. At this moment, the well-structured context is passed to the API of the LLM. This is where all the magic has to happen: based on its training material and given context, the LLM should give suggestions for code. This is represented with node F, colored in blue to indicate again that this is an important node.

  7. Generated Output The LLM returns its suggestions, and the user sees them inside the code editor. This could be code completions, code block suggestions, or even refactoring options, depending on how well the IDE understands the current context of your project. This is represented by node G.

References

https://www.youtube.com/watch?v=2ZpJXHiPwtQ “How useful is AI for programming? | Marc Andreessen and Lex Fridman” Lex Clips

https://medium.com/@anala007/hidden-dangers-of-using-cursor-ai-for-code-generation-what-every-developer-should-know-f4993c407b00

Use with Security

https://learning.oreilly.com/library/view/developing-cybersecurity-programs/9780138073992/ Developing Cybersecurity Programs and Policies in an AI-Driven World, 4th Edition By Omar Santos

https://learning.oreilly.com/live-events/-/0642572001334/ Cover of AI for Cybersecurity in 90 Minutes AI for Cybersecurity in 90 Minutes With Shaila Rana


References

https://www.ft.com/content/389e505c-a1cc-4176-a592-dd1d0fa171b8?utm_source=pocket-newtab-en-us

Infosys has a Springboard Digital Academy which provides a class on Prompt Engineering: https://infyspringboard.us.onwingspan.com/web/en/app/toc/lex_auth_013719953643773952304_shared/overview

https://www.unite.ai/zero-to-advanced-prompt-engineering-with-langchain-in-python/

VIDEO: “How to use ChatGPT to learn a language” (by English teacher learning Madarin)

https://arXiv_2307.10169_Challenges_and_Applications_of_Large_Language_Models.pdf

https://arxiv.org/pdf/2307.01850.pdf Risks of using synthetic data to train your model, (Self-Consuming Generative Models Go MAD)

https://www.onlinetutorials.org/teaching-academics/chatgpt4-for-medical-writers-and-editors/ ChatGPT4 for Medical Writers and Editors by https://www.linkedin.com/in/emmahittnichols/

https://thispersondoesnotexist.com/ displays full-screen photos using “Kerras, et al”

https://www.youtube.com/watch?v=4Icpq1vZkrw 99 GPT Prompts for Business Efficiency & Growth

https://learn.microsoft.com/en-us/training/modules/fundamentals-generative-ai/ Fundamentals of Generative AI

https://learn.microsoft.com/training/paths/introduction-generative-ai/ Microsoft Azure AI Fundamentals: Generative AI

https://ig.ft.com/generative-ai/ Generative AI exists because of the transformer

https://www.youtube.com/watch?v=4Qz4GfvjGLY With GenAI, an organization only needs a programmer and a prompt engineer. https://www.skool.com/new-society by David Ondrej of https://www.skool.com/new-society https://www.youtube.com/watch?v=9uQ-i3z_g0c

Pieter Levels @levelsio - single-employee millionaire

Devin r/webdev charge money on Upwork and Reddit

Microsoft’s AutoDev (see research paper) a team of agents: Tools library: file editing, retrieval, building, testing, git operations

Zapier Central

The future of AI is agentic. The basis of competition will be who can build better agents So building agents will be the most important skill

CrewAI

Claud 3 Haiku LLM costs $0.25 per million

emaggiori

Benchmarks

Neo4j Graph Databases

https://graphacademy.neo4j.com/ offers free courses on their Graph Vector databases for use in GenAI.

https://neo4j.com/graph-algorithms-book/?ref=blog Apache Spark and Neo4j

References

https://www.tiktok.com/@stevenouri/playlist/AI%20Generated-7351330477631998727 Steve Nouri - AI | ChatGPT linktr.ee/stevenouri

https://www.youtube.com/watch?v=T7TkxOMftz4 Jayme Edwards $7.99/month

GenAI CCS

https://store.certiport.com/msi-online-course-for-critical-career-skills-ccs-generative-ai-foundations-certification/p/12009958 $95 Critical Career Skills (CCS) Generative AI Foundations Certification on Pearson’s Certiport.com “The exam positions learners at the cutting edge of a rapidly emerging field. It provides a solid grounding in generative AI, enhancing their ability to adapt in a tech-driven job market.”

$235 with MSi Online Course

Johns Hopkins with Great Learning India

$2950 - $300 16 Weeks • Online Certificate Program in Applied Generative AI

Faculty members:

Great Learning is based in Bengaluru, Karnataka, India. VIDEO REVIEWS

Mentors: