From baf2364168b39ca4dae85f80ee5258e2c33c4589 Mon Sep 17 00:00:00 2001 From: gaikwadrahul8 <115997457+gaikwadrahul8@users.noreply.github.com> Date: Fri, 12 Apr 2024 23:22:06 +0530 Subject: [PATCH] Address tfjs-automl typos in documentation strings (#8219) --- tfjs-automl/README.md | 8 ++++---- tfjs-automl/src/test_browser.ts | 2 +- tfjs-automl/src/test_node.ts | 2 +- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/tfjs-automl/README.md b/tfjs-automl/README.md index 6dda1584419..1d3732713f9 100644 --- a/tfjs-automl/README.md +++ b/tfjs-automl/README.md @@ -48,10 +48,10 @@ const modelUrl = 'model.json'; // URL to the model.json file. const model = await automl.loadImageClassification(modelUrl); ``` -If you do not want (or cannot) load the model over HTTP you can also load the model separately and directly use the constuctor. +If you do not want (or cannot) load the model over HTTP you can also load the model separately and directly use the constructor. This is particularly relevant for __non-browser__ platforms. -The following psuedocode demonstrates this approach: +The following pseudocode demonstrates this approach: ```js import * as automl from '@tensorflow/tfjs-automl'; @@ -138,10 +138,10 @@ const modelUrl = 'model.json'; // URL to the model.json file. const model = await automl.loadObjectDetection(modelUrl); ``` -If you do not want (or cannot) load the model over HTTP you can also load the model separately and directly use the constuctor. +If you do not want (or cannot) load the model over HTTP you can also load the model separately and directly use the constructor. This is particularly relevant for __non-browser__ platforms. -The following psuedocode demonstrates this approach: +The following pseudocode demonstrates this approach: ```js import * as automl from '@tensorflow/tfjs-automl'; diff --git a/tfjs-automl/src/test_browser.ts b/tfjs-automl/src/test_browser.ts index 989bab4471a..44a65dbfb62 100644 --- a/tfjs-automl/src/test_browser.ts +++ b/tfjs-automl/src/test_browser.ts @@ -39,7 +39,7 @@ const testEnv = parseTestEnvFromKarmaFlags(__karma__.config.args, TEST_ENVS); if (testEnv != null) { setTestEnvs([testEnv]); } else { - // Run browser tests againts both the webgl backends. + // Run browser tests against both the webgl backends. setTestEnvs([ // WebGL. { diff --git a/tfjs-automl/src/test_node.ts b/tfjs-automl/src/test_node.ts index ee60399ea37..4d7b4208582 100644 --- a/tfjs-automl/src/test_node.ts +++ b/tfjs-automl/src/test_node.ts @@ -28,7 +28,7 @@ process.on('unhandledRejection', e => { throw e; }); -// Run node tests againts the cpu backend. +// Run node tests against the cpu backend. setTestEnvs([{name: 'node', backendName: 'cpu'}]); const runner = new jasmine();