-
Notifications
You must be signed in to change notification settings - Fork 2k
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Split analyze.js into two separate samples (#360)
- Loading branch information
Ace Nassri
authored
Apr 25, 2017
1 parent
1a91f3d
commit 0332a08
Showing
4 changed files
with
391 additions
and
1 deletion.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,291 @@ | ||
/** | ||
* Copyright 2017, Google, Inc. | ||
* Licensed under the Apache License, Version 2.0 (the "License"); | ||
* you may not use this file except in compliance with the License. | ||
* You may obtain a copy of the License at | ||
* | ||
* http://www.apache.org/licenses/LICENSE-2.0 | ||
* | ||
* Unless required by applicable law or agreed to in writing, software | ||
* distributed under the License is distributed on an "AS IS" BASIS, | ||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
* See the License for the specific language governing permissions and | ||
* limitations under the License. | ||
*/ | ||
|
||
'use strict'; | ||
|
||
function analyzeSentimentOfText (text) { | ||
// [START language_sentiment_string] | ||
// Imports the Google Cloud client library | ||
const Language = require('@google-cloud/language'); | ||
|
||
// Instantiates a client | ||
const language = Language(); | ||
|
||
// The text to analyze, e.g. "Hello, world!" | ||
// const text = 'Hello, world!'; | ||
|
||
// Instantiates a Document, representing the provided text | ||
const document = language.document({ content: text }); | ||
|
||
// Detects the sentiment of the document | ||
document.detectSentiment() | ||
.then((results) => { | ||
const sentiment = results[1].documentSentiment; | ||
console.log(`Document sentiment:`); | ||
console.log(` Score: ${sentiment.score}`); | ||
console.log(` Magnitude: ${sentiment.magnitude}`); | ||
|
||
const sentences = results[1].sentences; | ||
sentences.forEach((sentence) => { | ||
console.log(`Sentence: ${sentence.text.content}`); | ||
console.log(` Score: ${sentence.sentiment.score}`); | ||
console.log(` Magnitude: ${sentence.sentiment.magnitude}`); | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_sentiment_string] | ||
} | ||
|
||
function analyzeSentimentInFile (bucketName, fileName) { | ||
// [START language_sentiment_file] | ||
// Imports the Google Cloud client libraries | ||
const Language = require('@google-cloud/language'); | ||
const Storage = require('@google-cloud/storage'); | ||
|
||
// Instantiates the clients | ||
const language = Language(); | ||
const storage = Storage(); | ||
|
||
// The name of the bucket where the file resides, e.g. "my-bucket" | ||
// const bucketName = 'my-bucket'; | ||
|
||
// The name of the file to analyze, e.g. "file.txt" | ||
// const fileName = 'file.txt'; | ||
|
||
// Instantiates a Document, representing a text file in Cloud Storage | ||
const document = language.document({ | ||
// The Google Cloud Storage file | ||
content: storage.bucket(bucketName).file(fileName) | ||
}); | ||
|
||
// Detects the sentiment of the document | ||
document.detectSentiment() | ||
.then((results) => { | ||
const sentiment = results[1].documentSentiment; | ||
console.log(`Document sentiment:`); | ||
console.log(` Score: ${sentiment.score}`); | ||
console.log(` Magnitude: ${sentiment.magnitude}`); | ||
|
||
const sentences = results[1].sentences; | ||
sentences.forEach((sentence) => { | ||
console.log(`Sentence: ${sentence.text.content}`); | ||
console.log(` Score: ${sentence.sentiment.score}`); | ||
console.log(` Magnitude: ${sentence.sentiment.magnitude}`); | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_sentiment_file] | ||
} | ||
|
||
function analyzeEntitiesOfText (text) { | ||
// [START language_entities_string] | ||
// Imports the Google Cloud client library | ||
const Language = require('@google-cloud/language'); | ||
|
||
// Instantiates a client | ||
const language = Language(); | ||
|
||
// The text to analyze, e.g. "Hello, world!" | ||
// const text = 'Hello, world!'; | ||
|
||
// Instantiates a Document, representing the provided text | ||
const document = language.document({ content: text }); | ||
|
||
// Detects entities in the document | ||
document.detectEntities() | ||
.then((results) => { | ||
const entities = results[1].entities; | ||
|
||
console.log('Entities:'); | ||
entities.forEach((entity) => { | ||
console.log(entity.name); | ||
console.log(` - Type: ${entity.type}, Salience: ${entity.salience}`); | ||
if (entity.metadata && entity.metadata.wikipedia_url) { | ||
console.log(` - Wikipedia URL: ${entity.metadata.wikipedia_url}$`); | ||
} | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_entities_string] | ||
} | ||
|
||
function analyzeEntitiesInFile (bucketName, fileName) { | ||
// [START language_entities_file] | ||
// Imports the Google Cloud client libraries | ||
const Language = require('@google-cloud/language'); | ||
const Storage = require('@google-cloud/storage'); | ||
|
||
// Instantiates the clients | ||
const language = Language(); | ||
const storage = Storage(); | ||
|
||
// The name of the bucket where the file resides, e.g. "my-bucket" | ||
// const bucketName = 'my-bucket'; | ||
|
||
// The name of the file to analyze, e.g. "file.txt" | ||
// const fileName = 'file.txt'; | ||
|
||
// Instantiates a Document, representing a text file in Cloud Storage | ||
const document = language.document({ | ||
// The Google Cloud Storage file | ||
content: storage.bucket(bucketName).file(fileName) | ||
}); | ||
|
||
// Detects entities in the document | ||
document.detectEntities() | ||
.then((results) => { | ||
const entities = results[0]; | ||
|
||
console.log('Entities:'); | ||
entities.forEach((entity) => { | ||
console.log(entity.name); | ||
console.log(` - Type: ${entity.type}, Salience: ${entity.salience}`); | ||
if (entity.metadata && entity.metadata.wikipedia_url) { | ||
console.log(` - Wikipedia URL: ${entity.metadata.wikipedia_url}$`); | ||
} | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_entities_file] | ||
} | ||
|
||
function analyzeSyntaxOfText (text) { | ||
// [START language_syntax_string] | ||
// Imports the Google Cloud client library | ||
const Language = require('@google-cloud/language'); | ||
|
||
// Instantiates a client | ||
const language = Language(); | ||
|
||
// The text to analyze, e.g. "Hello, world!" | ||
// const text = 'Hello, world!'; | ||
|
||
// Instantiates a Document, representing the provided text | ||
const document = language.document({ content: text }); | ||
|
||
// Detects syntax in the document | ||
document.detectSyntax() | ||
.then((results) => { | ||
const syntax = results[0]; | ||
|
||
console.log('Parts of speech:'); | ||
syntax.forEach((part) => { | ||
console.log(`${part.partOfSpeech.tag}: ${part.text.content}`); | ||
console.log(`Morphology:`, part.partOfSpeech); | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_syntax_string] | ||
} | ||
|
||
function analyzeSyntaxInFile (bucketName, fileName) { | ||
// [START language_syntax_file] | ||
// Imports the Google Cloud client libraries | ||
const Language = require('@google-cloud/language'); | ||
const Storage = require('@google-cloud/storage'); | ||
|
||
// Instantiates the clients | ||
const language = Language(); | ||
const storage = Storage(); | ||
|
||
// The name of the bucket where the file resides, e.g. "my-bucket" | ||
// const bucketName = 'my-bucket'; | ||
|
||
// The name of the file to analyze, e.g. "file.txt" | ||
// const fileName = 'file.txt'; | ||
|
||
// Instantiates a Document, representing a text file in Cloud Storage | ||
const document = language.document({ | ||
// The Google Cloud Storage file | ||
content: storage.bucket(bucketName).file(fileName) | ||
}); | ||
|
||
// Detects syntax in the document | ||
document.detectSyntax() | ||
.then((results) => { | ||
const syntax = results[0]; | ||
|
||
console.log('Parts of speech:'); | ||
syntax.forEach((part) => { | ||
console.log(`${part.partOfSpeech.tag}: ${part.text.content}`); | ||
console.log(`Morphology:`, part.partOfSpeech); | ||
}); | ||
}) | ||
.catch((err) => { | ||
console.error('ERROR:', err); | ||
}); | ||
// [END language_syntax_file] | ||
} | ||
|
||
require(`yargs`) // eslint-disable-line | ||
.demand(1) | ||
.command( | ||
`sentiment-text <text>`, | ||
`Detects sentiment of a string.`, | ||
{}, | ||
(opts) => analyzeSentimentOfText(opts.text) | ||
) | ||
.command( | ||
`sentiment-file <bucketName> <fileName>`, | ||
`Detects sentiment in a file in Google Cloud Storage.`, | ||
{}, | ||
(opts) => analyzeSentimentInFile(opts.bucketName, opts.fileName) | ||
) | ||
.command( | ||
`entities-text <text>`, | ||
`Detects entities in a string.`, | ||
{}, | ||
(opts) => analyzeEntitiesOfText(opts.text) | ||
) | ||
.command( | ||
`entities-file <bucketName> <fileName>`, | ||
`Detects entities in a file in Google Cloud Storage.`, | ||
{}, | ||
(opts) => analyzeEntitiesInFile(opts.bucketName, opts.fileName) | ||
) | ||
.command( | ||
`syntax-text <text>`, | ||
`Detects syntax of a string.`, | ||
{}, | ||
(opts) => analyzeSyntaxOfText(opts.text) | ||
) | ||
.command( | ||
`syntax-file <bucketName> <fileName>`, | ||
`Detects syntax in a file in Google Cloud Storage.`, | ||
{}, | ||
(opts) => analyzeSyntaxInFile(opts.bucketName, opts.fileName) | ||
) | ||
.example(`node $0 sentiment-text "President Obama is speaking at the White House."`) | ||
.example(`node $0 sentiment-file my-bucket file.txt`, `Detects sentiment in gs://my-bucket/file.txt`) | ||
.example(`node $0 entities-text "President Obama is speaking at the White House."`) | ||
.example(`node $0 entities-file my-bucket file.txt`, `Detects entities in gs://my-bucket/file.txt`) | ||
.example(`node $0 syntax-text "President Obama is speaking at the White House."`) | ||
.example(`node $0 syntax-file my-bucket file.txt`, `Detects syntax in gs://my-bucket/file.txt`) | ||
.wrap(120) | ||
.recommendCommands() | ||
.epilogue(`For more information, see https://cloud.google.com/natural-language/docs`) | ||
.help() | ||
.strict() | ||
.argv; |
File renamed without changes.
99 changes: 99 additions & 0 deletions
99
cloud-language/snippets/system-test/analyze.v1beta1.test.js
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,99 @@ | ||
/** | ||
* Copyright 2017, Google, Inc. | ||
* Licensed under the Apache License, Version 2.0 (the "License"); | ||
* you may not use this file except in compliance with the License. | ||
* You may obtain a copy of the License at | ||
* | ||
* http://www.apache.org/licenses/LICENSE-2.0 | ||
* | ||
* Unless required by applicable law or agreed to in writing, software | ||
* distributed under the License is distributed on an "AS IS" BASIS, | ||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
* See the License for the specific language governing permissions and | ||
* limitations under the License. | ||
*/ | ||
|
||
'use strict'; | ||
|
||
const path = require(`path`); | ||
const storage = require(`@google-cloud/storage`)(); | ||
const test = require(`ava`); | ||
const tools = require(`@google-cloud/nodejs-repo-tools`); | ||
const uuid = require(`uuid`); | ||
|
||
const cmd = `node analyze.v1beta1.js`; | ||
const cwd = path.join(__dirname, `..`); | ||
const bucketName = `nodejs-docs-samples-test-${uuid.v4()}`; | ||
const fileName = `text.txt`; | ||
const localFilePath = path.join(__dirname, `../resources/text.txt`); | ||
const text = `President Obama is speaking at the White House.`; | ||
|
||
test.before(async () => { | ||
tools.checkCredentials(); | ||
const [bucket] = await storage.createBucket(bucketName); | ||
await bucket.upload(localFilePath); | ||
}); | ||
|
||
test.after.always(async () => { | ||
const bucket = storage.bucket(bucketName); | ||
await bucket.deleteFiles({ force: true }); | ||
await bucket.deleteFiles({ force: true }); // Try a second time... | ||
await bucket.delete(); | ||
}); | ||
|
||
test.beforeEach(tools.stubConsole); | ||
test.afterEach.always(tools.restoreConsole); | ||
|
||
test(`should analyze sentiment in text`, async (t) => { | ||
const output = await tools.runAsync(`${cmd} sentiment-text "${text}"`, cwd); | ||
t.true(output.includes(`Document sentiment:`)); | ||
t.true(output.includes(`Sentence: ${text}`)); | ||
t.true(output.includes(`Score: 0`)); | ||
t.true(output.includes(`Magnitude: 0`)); | ||
}); | ||
|
||
test(`should analyze sentiment in a file`, async (t) => { | ||
const output = await tools.runAsync(`${cmd} sentiment-file ${bucketName} ${fileName}`, cwd); | ||
t.true(output.includes(`Document sentiment:`)); | ||
t.true(output.includes(`Sentence: ${text}`)); | ||
t.true(output.includes(`Score: 0`)); | ||
t.true(output.includes(`Magnitude: 0`)); | ||
}); | ||
|
||
test(`should analyze entities in text`, async (t) => { | ||
const output = await tools.runAsync(`${cmd} entities-text "${text}"`, cwd); | ||
t.true(output.includes(`Obama`)); | ||
t.true(output.includes(`Type: PERSON`)); | ||
t.true(output.includes(`White House`)); | ||
t.true(output.includes(`Type: LOCATION`)); | ||
t.true(output.includes(`/wiki/Barack_Obama`)); | ||
}); | ||
|
||
test('should analyze entities in a file', async (t) => { | ||
const output = await tools.runAsync(`${cmd} entities-file ${bucketName} ${fileName}`, cwd); | ||
t.true(output.includes(`Entities:`)); | ||
t.true(output.includes(`Obama`)); | ||
t.true(output.includes(`Type: PERSON`)); | ||
t.true(output.includes(`White House`)); | ||
t.true(output.includes(`Type: LOCATION`)); | ||
t.true(output.includes(`/wiki/Barack_Obama`)); | ||
}); | ||
|
||
test(`should analyze syntax in text`, async (t) => { | ||
const output = await tools.runAsync(`${cmd} syntax-text "${text}"`, cwd); | ||
t.true(output.includes(`Parts of speech:`)); | ||
t.true(output.includes(`NOUN:`)); | ||
t.true(output.includes(`President`)); | ||
t.true(output.includes(`Obama`)); | ||
t.true(output.includes(`Morphology:`)); | ||
t.true(output.includes(`tag: 'NOUN'`)); | ||
}); | ||
|
||
test('should analyze syntax in a file', async (t) => { | ||
const output = await tools.runAsync(`${cmd} syntax-file ${bucketName} ${fileName}`, cwd); | ||
t.true(output.includes(`NOUN:`)); | ||
t.true(output.includes(`President`)); | ||
t.true(output.includes(`Obama`)); | ||
t.true(output.includes(`Morphology:`)); | ||
t.true(output.includes(`tag: 'NOUN'`)); | ||
}); |
Oops, something went wrong.