From fecae45b46cd6139bbf1574d86bedb77172e0da6 Mon Sep 17 00:00:00 2001 From: ErdemOzgen <14043035+ErdemOzgen@users.noreply.github.com> Date: Sat, 5 Oct 2024 00:24:35 +0300 Subject: [PATCH] LLM cyber book --- ...ecurity Threats, Exposure and Mitigation Springer.md | 9 +++++++++ content/Article&Books/books/Book Index.md | 2 +- 2 files changed, 10 insertions(+), 1 deletion(-) create mode 100644 content/AI&DATA/Large Language Models in Cybersecurity Threats, Exposure and Mitigation Springer.md diff --git a/content/AI&DATA/Large Language Models in Cybersecurity Threats, Exposure and Mitigation Springer.md b/content/AI&DATA/Large Language Models in Cybersecurity Threats, Exposure and Mitigation Springer.md new file mode 100644 index 000000000..18f3adeb3 --- /dev/null +++ b/content/AI&DATA/Large Language Models in Cybersecurity Threats, Exposure and Mitigation Springer.md @@ -0,0 +1,9 @@ + + + +Such a trade-off is known as scaling laws—an optimal relationship between the model size, dataset size, and computational power investment required to achieve the best model performance while minimizing computational expenses. There are currently two schools of thought on what this scaling law is. + +[[Chinchilla scaling law]] + +Perhaps the most known example is the so-called “[[Chain-of-thought]]” prompting, published in early 2022 . The trick in that technique is to prime the generation of more high-quality reasoning-like utterances by providing an example of a solution to a reasoning problem that makes all steps explicit rather than directly outputting the answer, allowing the attention maps to better identify and articulate in a more plausible manner predicates involved in reasoning. Shortly after, an even more powerful, zero-shot approach was identified when a prompt requesting reasoning was terminated with a “let us reason step-by-step”, restraining the generation space to educational examples of reasoning present in the training dataset, at which point their structure allowed the attention maps to operate in the case similar to the few-shot approach previously presented. + diff --git a/content/Article&Books/books/Book Index.md b/content/Article&Books/books/Book Index.md index 5736ddd69..d25ffac1d 100644 --- a/content/Article&Books/books/Book Index.md +++ b/content/Article&Books/books/Book Index.md @@ -3,7 +3,7 @@ # AI Books [[Hands on Machine Learning]] - +[[Large Language Models in Cybersecurity Threats, Exposure and Mitigation Springer ]] # Software Enginnering