You're Invited:Meet the Socket Team at RSAC and BSidesSF 2026, March 23–26.RSVP
Socket
Book a DemoSign in
Socket

@ai-sdk/google

Package Overview
Dependencies
Maintainers
3
Versions
408
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

@ai-sdk/google - npm Package Compare versions

Comparing version
4.0.0-beta.8
to
4.0.0-beta.9
+6
-0
CHANGELOG.md
# @ai-sdk/google
## 4.0.0-beta.9
### Patch Changes
- e59176c: chore(google): update v3 specs to v4
## 4.0.0-beta.8

@@ -4,0 +10,0 @@

+13
-13
import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
import { ProviderV3, LanguageModelV3, ImageModelV3, EmbeddingModelV3, Experimental_VideoModelV3 } from '@ai-sdk/provider';
import { ProviderV4, LanguageModelV4, ImageModelV4, EmbeddingModelV4, Experimental_VideoModelV4 } from '@ai-sdk/provider';

@@ -305,38 +305,38 @@ declare const googleErrorDataSchema: _ai_sdk_provider_utils.LazySchema<{

interface GoogleGenerativeAIProvider extends ProviderV3 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
interface GoogleGenerativeAIProvider extends ProviderV4 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
/**
* Creates a model for image generation.
*/
image(modelId: GoogleGenerativeAIImageModelId, settings?: GoogleGenerativeAIImageSettings): ImageModelV3;
image(modelId: GoogleGenerativeAIImageModelId, settings?: GoogleGenerativeAIImageSettings): ImageModelV4;
/**
* @deprecated Use `chat()` instead.
*/
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
/**
* Creates a model for text embeddings.
*/
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* Creates a model for text embeddings.
*/
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* @deprecated Use `embedding` instead.
*/
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* @deprecated Use `embeddingModel` instead.
*/
textEmbeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
textEmbeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* Creates a model for video generation.
*/
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV3;
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV4;
/**
* Creates a model for video generation.
*/
videoModel(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV3;
videoModel(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV4;
tools: typeof googleTools;

@@ -343,0 +343,0 @@ }

import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
import { ProviderV3, LanguageModelV3, ImageModelV3, EmbeddingModelV3, Experimental_VideoModelV3 } from '@ai-sdk/provider';
import { ProviderV4, LanguageModelV4, ImageModelV4, EmbeddingModelV4, Experimental_VideoModelV4 } from '@ai-sdk/provider';

@@ -305,38 +305,38 @@ declare const googleErrorDataSchema: _ai_sdk_provider_utils.LazySchema<{

interface GoogleGenerativeAIProvider extends ProviderV3 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
interface GoogleGenerativeAIProvider extends ProviderV4 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
/**
* Creates a model for image generation.
*/
image(modelId: GoogleGenerativeAIImageModelId, settings?: GoogleGenerativeAIImageSettings): ImageModelV3;
image(modelId: GoogleGenerativeAIImageModelId, settings?: GoogleGenerativeAIImageSettings): ImageModelV4;
/**
* @deprecated Use `chat()` instead.
*/
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
/**
* Creates a model for text embeddings.
*/
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* Creates a model for text embeddings.
*/
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* @deprecated Use `embedding` instead.
*/
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* @deprecated Use `embeddingModel` instead.
*/
textEmbeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
textEmbeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;
/**
* Creates a model for video generation.
*/
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV3;
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV4;
/**
* Creates a model for video generation.
*/
videoModel(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV3;
videoModel(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV4;
tools: typeof googleTools;

@@ -343,0 +343,0 @@ }

import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
import { Resolvable, FetchFunction, InferSchema } from '@ai-sdk/provider-utils';
import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult } from '@ai-sdk/provider';
import { LanguageModelV4, LanguageModelV4CallOptions, LanguageModelV4GenerateResult, LanguageModelV4StreamResult } from '@ai-sdk/provider';
import { z } from 'zod/v4';

@@ -17,6 +17,6 @@

*/
supportedUrls?: () => LanguageModelV3['supportedUrls'];
supportedUrls?: () => LanguageModelV4['supportedUrls'];
};
declare class GoogleGenerativeAILanguageModel implements LanguageModelV3 {
readonly specificationVersion = "v3";
declare class GoogleGenerativeAILanguageModel implements LanguageModelV4 {
readonly specificationVersion = "v4";
readonly modelId: GoogleGenerativeAIModelId;

@@ -29,4 +29,4 @@ private readonly config;

private getArgs;
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
doGenerate(options: LanguageModelV4CallOptions): Promise<LanguageModelV4GenerateResult>;
doStream(options: LanguageModelV4CallOptions): Promise<LanguageModelV4StreamResult>;
}

@@ -33,0 +33,0 @@ declare const getGroundingMetadataSchema: () => z.ZodObject<{

import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
import { Resolvable, FetchFunction, InferSchema } from '@ai-sdk/provider-utils';
import { LanguageModelV3, LanguageModelV3CallOptions, LanguageModelV3GenerateResult, LanguageModelV3StreamResult } from '@ai-sdk/provider';
import { LanguageModelV4, LanguageModelV4CallOptions, LanguageModelV4GenerateResult, LanguageModelV4StreamResult } from '@ai-sdk/provider';
import { z } from 'zod/v4';

@@ -17,6 +17,6 @@

*/
supportedUrls?: () => LanguageModelV3['supportedUrls'];
supportedUrls?: () => LanguageModelV4['supportedUrls'];
};
declare class GoogleGenerativeAILanguageModel implements LanguageModelV3 {
readonly specificationVersion = "v3";
declare class GoogleGenerativeAILanguageModel implements LanguageModelV4 {
readonly specificationVersion = "v4";
readonly modelId: GoogleGenerativeAIModelId;

@@ -29,4 +29,4 @@ private readonly config;

private getArgs;
doGenerate(options: LanguageModelV3CallOptions): Promise<LanguageModelV3GenerateResult>;
doStream(options: LanguageModelV3CallOptions): Promise<LanguageModelV3StreamResult>;
doGenerate(options: LanguageModelV4CallOptions): Promise<LanguageModelV4GenerateResult>;
doStream(options: LanguageModelV4CallOptions): Promise<LanguageModelV4StreamResult>;
}

@@ -33,0 +33,0 @@ declare const getGroundingMetadataSchema: () => z.ZodObject<{

@@ -747,3 +747,3 @@ "use strict";

constructor(modelId, config) {
this.specificationVersion = "v3";
this.specificationVersion = "v4";
var _a;

@@ -750,0 +750,0 @@ this.modelId = modelId;

@@ -736,3 +736,3 @@ // src/google-generative-ai-language-model.ts

constructor(modelId, config) {
this.specificationVersion = "v3";
this.specificationVersion = "v4";
var _a;

@@ -739,0 +739,0 @@ this.modelId = modelId;

{
"name": "@ai-sdk/google",
"version": "4.0.0-beta.8",
"version": "4.0.0-beta.9",
"license": "Apache-2.0",

@@ -5,0 +5,0 @@ "sideEffects": false,

@@ -1,2 +0,2 @@

import { LanguageModelV3Usage } from '@ai-sdk/provider';
import { LanguageModelV4Usage } from '@ai-sdk/provider';

@@ -14,3 +14,3 @@ export type GoogleGenerativeAIUsageMetadata = {

usage: GoogleGenerativeAIUsageMetadata | undefined | null,
): LanguageModelV3Usage {
): LanguageModelV4Usage {
if (usage == null) {

@@ -17,0 +17,0 @@ return {

import {
LanguageModelV3Prompt,
LanguageModelV4Prompt,
UnsupportedFunctionalityError,

@@ -13,3 +13,3 @@ } from '@ai-sdk/provider';

export function convertToGoogleGenerativeAIMessages(
prompt: LanguageModelV3Prompt,
prompt: LanguageModelV4Prompt,
options?: { isGemmaModel?: boolean; providerOptionsName?: string },

@@ -16,0 +16,0 @@ ): GoogleGenerativeAIPrompt {

import {
EmbeddingModelV3,
EmbeddingModelV4,
TooManyEmbeddingValuesForCallError,

@@ -29,4 +29,4 @@ } from '@ai-sdk/provider';

export class GoogleGenerativeAIEmbeddingModel implements EmbeddingModelV3 {
readonly specificationVersion = 'v3';
export class GoogleGenerativeAIEmbeddingModel implements EmbeddingModelV4 {
readonly specificationVersion = 'v4';
readonly modelId: GoogleGenerativeAIEmbeddingModelId;

@@ -54,4 +54,4 @@ readonly maxEmbeddingsPerCall = 2048;

providerOptions,
}: Parameters<EmbeddingModelV3['doEmbed']>[0]): Promise<
Awaited<ReturnType<EmbeddingModelV3['doEmbed']>>
}: Parameters<EmbeddingModelV4['doEmbed']>[0]): Promise<
Awaited<ReturnType<EmbeddingModelV4['doEmbed']>>
> {

@@ -58,0 +58,0 @@ // Parse provider options

import {
ImageModelV3,
LanguageModelV3Prompt,
SharedV3Warning,
ImageModelV4,
LanguageModelV4Prompt,
SharedV4Warning,
} from '@ai-sdk/provider';

@@ -40,4 +40,4 @@ import {

export class GoogleGenerativeAIImageModel implements ImageModelV3 {
readonly specificationVersion = 'v3';
export class GoogleGenerativeAIImageModel implements ImageModelV4 {
readonly specificationVersion = 'v4';

@@ -67,4 +67,4 @@ get maxImagesPerCall(): number {

async doGenerate(
options: Parameters<ImageModelV3['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>> {
options: Parameters<ImageModelV4['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV4['doGenerate']>>> {
// Gemini image models use the language model API internally

@@ -78,4 +78,4 @@ if (isGeminiModel(this.modelId)) {

private async doGenerateImagen(
options: Parameters<ImageModelV3['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>> {
options: Parameters<ImageModelV4['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV4['doGenerate']>>> {
const {

@@ -93,3 +93,3 @@ prompt,

} = options;
const warnings: Array<SharedV3Warning> = [];
const warnings: Array<SharedV4Warning> = [];

@@ -188,4 +188,4 @@ // Imagen API endpoints do not support image editing

private async doGenerateGemini(
options: Parameters<ImageModelV3['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>> {
options: Parameters<ImageModelV4['doGenerate']>[0],
): Promise<Awaited<ReturnType<ImageModelV4['doGenerate']>>> {
const {

@@ -203,3 +203,3 @@ prompt,

} = options;
const warnings: Array<SharedV3Warning> = [];
const warnings: Array<SharedV4Warning> = [];

@@ -262,3 +262,3 @@ // Gemini does not support mask-based inpainting

const languageModelPrompt: LanguageModelV3Prompt = [
const languageModelPrompt: LanguageModelV4Prompt = [
{ role: 'user', content: userContent },

@@ -265,0 +265,0 @@ ];

import {
LanguageModelV3,
LanguageModelV3CallOptions,
LanguageModelV3Content,
LanguageModelV3FinishReason,
LanguageModelV3GenerateResult,
LanguageModelV3Source,
LanguageModelV3StreamPart,
LanguageModelV3StreamResult,
SharedV3ProviderMetadata,
SharedV3Warning,
LanguageModelV4,
LanguageModelV4CallOptions,
LanguageModelV4Content,
LanguageModelV4FinishReason,
LanguageModelV4GenerateResult,
LanguageModelV4Source,
LanguageModelV4StreamPart,
LanguageModelV4StreamResult,
SharedV4ProviderMetadata,
SharedV4Warning,
} from '@ai-sdk/provider';

@@ -58,7 +58,7 @@ import {

*/
supportedUrls?: () => LanguageModelV3['supportedUrls'];
supportedUrls?: () => LanguageModelV4['supportedUrls'];
};
export class GoogleGenerativeAILanguageModel implements LanguageModelV3 {
readonly specificationVersion = 'v3';
export class GoogleGenerativeAILanguageModel implements LanguageModelV4 {
readonly specificationVersion = 'v4';

@@ -101,4 +101,4 @@ readonly modelId: GoogleGenerativeAIModelId;

providerOptions,
}: LanguageModelV3CallOptions) {
const warnings: SharedV3Warning[] = [];
}: LanguageModelV4CallOptions) {
const warnings: SharedV4Warning[] = [];

@@ -214,4 +214,4 @@ const providerOptionsName = this.config.provider.includes('vertex')

async doGenerate(
options: LanguageModelV3CallOptions,
): Promise<LanguageModelV3GenerateResult> {
options: LanguageModelV4CallOptions,
): Promise<LanguageModelV4GenerateResult> {
const { args, warnings, providerOptionsName } = await this.getArgs(options);

@@ -241,3 +241,3 @@

const candidate = response.candidates[0];
const content: Array<LanguageModelV3Content> = [];
const content: Array<LanguageModelV4Content> = [];

@@ -378,4 +378,4 @@ // map ordered parts to content:

async doStream(
options: LanguageModelV3CallOptions,
): Promise<LanguageModelV3StreamResult> {
options: LanguageModelV4CallOptions,
): Promise<LanguageModelV4StreamResult> {
const { args, warnings, providerOptionsName } = await this.getArgs(options);

@@ -400,3 +400,3 @@

let finishReason: LanguageModelV3FinishReason = {
let finishReason: LanguageModelV4FinishReason = {
unified: 'other',

@@ -406,3 +406,3 @@ raw: undefined,

let usage: GoogleGenerativeAIUsageMetadata | undefined = undefined;
let providerMetadata: SharedV3ProviderMetadata | undefined = undefined;
let providerMetadata: SharedV4ProviderMetadata | undefined = undefined;
let lastGroundingMetadata: GroundingMetadataSchema | null = null;

@@ -428,3 +428,3 @@ let lastUrlContextMetadata: UrlContextMetadataSchema | null = null;

ParseResult<ChunkSchema>,
LanguageModelV3StreamPart
LanguageModelV4StreamPart
>({

@@ -766,3 +766,3 @@ start(controller) {

generateId: () => string;
}): undefined | LanguageModelV3Source[] {
}): undefined | LanguageModelV4Source[] {
if (!groundingMetadata?.groundingChunks) {

@@ -772,3 +772,3 @@ return undefined;

const sources: LanguageModelV3Source[] = [];
const sources: LanguageModelV4Source[] = [];

@@ -775,0 +775,0 @@ for (const chunk of groundingMetadata.groundingChunks) {

import {
AISDKError,
type Experimental_VideoModelV3,
type SharedV3Warning,
type Experimental_VideoModelV4,
type SharedV4Warning,
} from '@ai-sdk/provider';

@@ -53,4 +53,4 @@ import {

export class GoogleGenerativeAIVideoModel implements Experimental_VideoModelV3 {
readonly specificationVersion = 'v3';
export class GoogleGenerativeAIVideoModel implements Experimental_VideoModelV4 {
readonly specificationVersion = 'v4';

@@ -72,6 +72,6 @@ get provider(): string {

async doGenerate(
options: Parameters<Experimental_VideoModelV3['doGenerate']>[0],
): Promise<Awaited<ReturnType<Experimental_VideoModelV3['doGenerate']>>> {
options: Parameters<Experimental_VideoModelV4['doGenerate']>[0],
): Promise<Awaited<ReturnType<Experimental_VideoModelV4['doGenerate']>>> {
const currentDate = this.config._internal?.currentDate?.() ?? new Date();
const warnings: SharedV3Warning[] = [];
const warnings: SharedV4Warning[] = [];

@@ -78,0 +78,0 @@ const googleOptions = (await parseProviderOptions({

import {
LanguageModelV3CallOptions,
SharedV3Warning,
LanguageModelV4CallOptions,
SharedV4Warning,
UnsupportedFunctionalityError,

@@ -14,4 +14,4 @@ } from '@ai-sdk/provider';

}: {
tools: LanguageModelV3CallOptions['tools'];
toolChoice?: LanguageModelV3CallOptions['toolChoice'];
tools: LanguageModelV4CallOptions['tools'];
toolChoice?: LanguageModelV4CallOptions['toolChoice'];
modelId: GoogleGenerativeAIModelId;

@@ -39,3 +39,3 @@ }): {

};
toolWarnings: SharedV3Warning[];
toolWarnings: SharedV4Warning[];
} {

@@ -45,3 +45,3 @@ // when the tools array is empty, change it to undefined to prevent errors:

const toolWarnings: SharedV3Warning[] = [];
const toolWarnings: SharedV4Warning[] = [];

@@ -48,0 +48,0 @@ const isLatest = (

import {
EmbeddingModelV3,
Experimental_VideoModelV3,
ImageModelV3,
LanguageModelV3,
ProviderV3,
EmbeddingModelV4,
Experimental_VideoModelV4,
ImageModelV4,
LanguageModelV4,
ProviderV4,
} from '@ai-sdk/provider';

@@ -30,8 +30,8 @@ import {

export interface GoogleGenerativeAIProvider extends ProviderV3 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
export interface GoogleGenerativeAIProvider extends ProviderV4 {
(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
languageModel(modelId: GoogleGenerativeAIModelId): LanguageModelV4;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
chat(modelId: GoogleGenerativeAIModelId): LanguageModelV4;

@@ -44,3 +44,3 @@ /**

settings?: GoogleGenerativeAIImageSettings,
): ImageModelV3;
): ImageModelV4;

@@ -50,3 +50,3 @@ /**

*/
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV3;
generativeAI(modelId: GoogleGenerativeAIModelId): LanguageModelV4;

@@ -56,3 +56,3 @@ /**

*/
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;

@@ -62,3 +62,3 @@ /**

*/
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
embeddingModel(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;

@@ -68,3 +68,3 @@ /**

*/
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV3;
textEmbedding(modelId: GoogleGenerativeAIEmbeddingModelId): EmbeddingModelV4;

@@ -76,3 +76,3 @@ /**

modelId: GoogleGenerativeAIEmbeddingModelId,
): EmbeddingModelV3;
): EmbeddingModelV4;

@@ -82,3 +82,3 @@ /**

*/
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV3;
video(modelId: GoogleGenerativeAIVideoModelId): Experimental_VideoModelV4;

@@ -90,3 +90,3 @@ /**

modelId: GoogleGenerativeAIVideoModelId,
): Experimental_VideoModelV3;
): Experimental_VideoModelV4;

@@ -216,3 +216,3 @@ tools: typeof googleTools;

provider.specificationVersion = 'v3' as const;
provider.specificationVersion = 'v4' as const;
provider.languageModel = createChatModel;

@@ -219,0 +219,0 @@ provider.chat = createChatModel;

@@ -1,2 +0,2 @@

import { LanguageModelV3FinishReason } from '@ai-sdk/provider';
import { LanguageModelV4FinishReason } from '@ai-sdk/provider';

@@ -9,3 +9,3 @@ export function mapGoogleGenerativeAIFinishReason({

hasToolCalls: boolean;
}): LanguageModelV3FinishReason['unified'] {
}): LanguageModelV4FinishReason['unified'] {
switch (finishReason) {

@@ -12,0 +12,0 @@ case 'STOP':

Sorry, the diff of this file is too big to display

Sorry, the diff of this file is too big to display

Sorry, the diff of this file is too big to display

Sorry, the diff of this file is too big to display

Sorry, the diff of this file is too big to display

Sorry, the diff of this file is too big to display