Security News
PyPI’s New Archival Feature Closes a Major Security Gap
PyPI now allows maintainers to archive projects, improving security and helping users make informed decisions about their dependencies.
@data-client/normalizr
Advanced tools
Normalizes and denormalizes JSON according to schema for Redux and Flux applications
Install from the NPM repository using yarn or npm:
yarn add @data-client/normalizr
npm install --save @data-client/normalizr
Many APIs, public or not, return JSON data that has deeply nested objects. Using data in this kind of structure is often very difficult for JavaScript applications, especially those using Flux or Redux.
Normalizr is a small, but powerful utility for taking JSON with a schema definition and returning nested entities with their IDs, gathered in dictionaries.
Consider a typical blog post. The API response for a single post might look something like this:
{
"id": "123",
"author": {
"id": "1",
"name": "Paul"
},
"title": "My awesome blog post",
"comments": [
{
"id": "324",
"createdAt": "2013-05-29T00:00:00-04:00",
"commenter": {
"id": "2",
"name": "Nicole"
}
},
{
"id": "544",
"createdAt": "2013-05-30T00:00:00-04:00",
"commenter": {
"id": "1",
"name": "Paul"
}
}
]
}
We have two nested entity types within our article
: users
and comments
. Using various schema, we can normalize all three entity types down:
import { schema, Entity } from '@data-client/endpoint';
import { Temporal } from '@js-temporal/polyfill';
// Define a users schema
class User extends Entity {
pk() {
return this.id;
}
}
// Define your comments schema
class Comment extends Entity {
pk() {
return this.id;
}
static schema = {
commenter: User,
createdAt: Temporal.Instant.from,
};
}
// Define your article
class Article extends Entity {
pk() {
return this.id;
}
static schema = {
author: User,
comments: [Comment],
};
}
import { normalize } from '@data-client/normalizr';
const args = [{ id: '123' }];
const normalizedData = normalize(Article, originalData, args);
Now, normalizedData
will create a single serializable source of truth for all entities:
{
result: "123",
entities: {
articles: {
"123": {
id: "123",
author: "1",
title: "My awesome blog post",
comments: [ "324", "544" ]
}
},
users: {
"1": { "id": "1", "name": "Paul" },
"2": { "id": "2", "name": "Nicole" }
},
comments: {
"324": {
id: "324",
createdAt: "2013-05-29T00:00:00-04:00",
commenter: "2"
},
"544": {
id: "544",
createdAt: "2013-05-30T00:00:00-04:00",
commenter: "1"
}
}
},
}
normalizedData
can be placed in any flux store as the single source of truth for this data.
Accessing the store can then be done using flux selectors
by denormalizing
:
import { denormalize } from '@data-client/normalizr';
const denormalizedData = denormalize(
Article,
normalizedData.result,
normalizedData.entities,
args,
);
Now, denormalizedData
will instantiate the classes, ensuring all instances of the same member (like Paul
) are referentially equal:
Article {
id: '123',
title: 'My awesome blog post',
author: User { id: '1', name: 'Paul' },
comments: [
Comment {
id: '324',
createdAt: Instant [Temporal.Instant] {},
commenter: [User { id: '2', name: 'Nicole' }]
},
Comment {
id: '544',
createdAt: Instant [Temporal.Instant] {},
commenter: [User { id: '1', name: 'Paul' }]
}
]
}
MemoCache
is a singleton that can be used to maintain referential equality between calls as well
as potentially improved performance by 2000%. All three methods are memoized.
import { MemoCache } from '@data-client/normalizr';
// you can construct a new memo anytime you want to reset the cache
const memo = new MemoCache();
const { data, paths } = memo.denormalize(schema, input, state.entities, args);
const data = memo.query(schema, args, state.entities, state.indexes);
function query(schema, args, state, key) {
const queryKey = memo.buildQueryKey(
schema,
args,
state.entities,
state.indexes,
key,
);
const { data } = this.denormalize(schema, queryKey, state.entities, args);
return typeof data === 'symbol' ? undefined : (data as any);
}
memo.denormalize()
is just like denormalize() above but includes paths
as part of the return value. paths
is an Array of paths of all entities included in the result.
memo.query()
allows denormalizing without a normalized input. See Queryable for more info.
memo.buildQueryKey()
builds the input used to denormalize for query()
. This is exposed
to allow greater flexibility in its usage.
Performance compared to normalizr package (higher is better):
no cache | with cache | |
---|---|---|
normalize (long) | 119% | 119% |
denormalize (long) | 158% | 1,262% |
denormalize (short) | 676% | 2,367% |
Normalizr Client is based on Normalizr - originally created by Dan Abramov and inspired by a conversation with Jing Chen. Since v3, it was completely rewritten and maintained by Paul Armstrong.
Normalizr Client was rewritten and maintained by Normalizr contributor Nathaniel Tucker. It has also received much help, enthusiasm, and contributions from community members.
FAQs
Normalizes and denormalizes JSON according to schema for Redux and Flux applications
The npm package @data-client/normalizr receives a total of 543 weekly downloads. As such, @data-client/normalizr popularity was classified as not popular.
We found that @data-client/normalizr demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 0 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
PyPI now allows maintainers to archive projects, improving security and helping users make informed decisions about their dependencies.
Research
Security News
Malicious npm package postcss-optimizer delivers BeaverTail malware, targeting developer systems; similarities to past campaigns suggest a North Korean connection.
Security News
CISA's KEV data is now on GitHub, offering easier access, API integration, commit history tracking, and automated updates for security teams and researchers.