@promptbook/pdf
Advanced tools
Comparing version 0.72.0-33 to 0.72.0-34
@@ -7,3 +7,3 @@ import type { string_executable_path } from '../../types/typeAliases'; | ||
*/ | ||
export declare function locateLibreoffice(): Promise<string_executable_path>; | ||
export declare function locateLibreoffice(): Promise<string_executable_path | null>; | ||
/** | ||
@@ -10,0 +10,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` OR `@promptbook/legacy-documents` |
@@ -7,3 +7,3 @@ import type { string_executable_path } from '../../types/typeAliases'; | ||
*/ | ||
export declare function locatePandoc(): Promise<string_executable_path>; | ||
export declare function locatePandoc(): Promise<string_executable_path | null>; | ||
/** | ||
@@ -10,0 +10,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` OR `@promptbook/documents` |
@@ -29,3 +29,3 @@ import type { RequireAtLeastOne } from 'type-fest'; | ||
*/ | ||
export declare function locateApp(options: RequireAtLeastOne<LocateAppOptions, 'linuxWhich' | 'windowsSuffix' | 'macOsName'>): Promise<string_executable_path>; | ||
export declare function locateApp(options: RequireAtLeastOne<LocateAppOptions, 'linuxWhich' | 'windowsSuffix' | 'macOsName'>): Promise<string_executable_path | null>; | ||
/** | ||
@@ -32,0 +32,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` |
@@ -0,3 +1,3 @@ | ||
import type { string_executable_path } from '../../types/typeAliases'; | ||
import type { LocateAppOptions } from '../locateApp'; | ||
import type { string_executable_path } from '../../types/typeAliases'; | ||
/** | ||
@@ -8,3 +8,3 @@ * @@@ | ||
*/ | ||
export declare function locateAppOnLinux({ appName, linuxWhich, }: Pick<Required<LocateAppOptions>, 'appName' | 'linuxWhich'>): Promise<string_executable_path>; | ||
export declare function locateAppOnLinux({ appName, linuxWhich, }: Pick<Required<LocateAppOptions>, 'appName' | 'linuxWhich'>): Promise<string_executable_path | null>; | ||
/** | ||
@@ -11,0 +11,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` |
@@ -8,3 +8,3 @@ import type { string_executable_path } from '../../types/typeAliases'; | ||
*/ | ||
export declare function locateAppOnMacOs({ appName, macOsName, }: Pick<Required<LocateAppOptions>, 'appName' | 'macOsName'>): Promise<string_executable_path>; | ||
export declare function locateAppOnMacOs({ appName, macOsName, }: Pick<Required<LocateAppOptions>, 'appName' | 'macOsName'>): Promise<string_executable_path | null>; | ||
/** | ||
@@ -11,0 +11,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` |
@@ -8,3 +8,3 @@ import type { string_executable_path } from '../../types/typeAliases'; | ||
*/ | ||
export declare function locateAppOnWindows({ appName, windowsSuffix, }: Pick<Required<LocateAppOptions>, 'appName' | 'windowsSuffix'>): Promise<string_executable_path>; | ||
export declare function locateAppOnWindows({ appName, windowsSuffix, }: Pick<Required<LocateAppOptions>, 'appName' | 'windowsSuffix'>): Promise<string_executable_path | null>; | ||
/** | ||
@@ -11,0 +11,0 @@ * TODO: [๐ง ][โฟ] Maybe export through `@promptbook/node` |
{ | ||
"name": "@promptbook/pdf", | ||
"version": "0.72.0-33", | ||
"version": "0.72.0-34", | ||
"description": "Supercharge your use of large language models", | ||
@@ -55,3 +55,3 @@ "private": false, | ||
"peerDependencies": { | ||
"@promptbook/core": "0.72.0-33" | ||
"@promptbook/core": "0.72.0-34" | ||
}, | ||
@@ -58,0 +58,0 @@ "dependencies": { |
@@ -57,5 +57,5 @@ <!-- โ ๏ธ WARNING: This code has been generated so that any manual changes will be overwritten --> | ||
If you have a simple, single prompt for ChatGPT, GPT-4, Anthropic Claude, Google Gemini, Llama 2, or whatever, it doesn't matter how you integrate it. Whether it's calling a REST API directly, using the SDK, hardcoding the prompt into the source code, or importing a text file, the process remains the same. | ||
If you have a simple, single prompt for ChatGPT, GPT-4, Anthropic Claude, Google Gemini, Llama 3, or whatever, it doesn't matter how you integrate it. Whether it's calling a REST API directly, using the SDK, hardcoding the prompt into the source code, or importing a text file, the process remains the same. | ||
But often you will struggle with the limitations of LLMs, such as hallucinations, off-topic responses, poor quality output, language drift, word repetition repetition repetition repetition or misuse, lack of context, or just plain w๐๐ขrd responses. When this happens, you generally have three options: | ||
But often you will struggle with the **limitations of LLMs**, such as **hallucinations, off-topic responses, poor quality output, language and prompt drift, word repetition repetition repetition repetition or misuse, lack of context, or just plain w๐๐ขrd responses**. When this happens, you generally have three options: | ||
@@ -66,3 +66,3 @@ 1. **Fine-tune** the model to your specifications or even train your own. | ||
In all of these situations, but especially in 3., the Promptbook library can make your life easier. | ||
In all of these situations, but especially in 3., the **โจ Promptbook can make your life waaaaaaaaaay easier**. | ||
@@ -69,0 +69,0 @@ - [**Separates concerns**](https://github.com/webgptorg/promptbook/discussions/32) between prompt-engineer and programmer, between code files and prompt files, and between prompts and their execution logic. |
Sorry, the diff of this file is too big to display
Sorry, the diff of this file is too big to display
1332791