Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
23,447
11,965,044,260
IssuesEvent
2020-04-05 21:49:35
libass/libass
https://api.github.com/repos/libass/libass
closed
cache grows too large
performance regression
This eats memory at the rate of 1 or 2 MB per second: `mpv /dev/zero --demuxer=rawvideo --osd-fractions --osd-level=3` It didn't happen with the previous libass release.
True
cache grows too large - This eats memory at the rate of 1 or 2 MB per second: `mpv /dev/zero --demuxer=rawvideo --osd-fractions --osd-level=3` It didn't happen with the previous libass release.
non_priority
cache grows too large this eats memory at the rate of or mb per second mpv dev zero demuxer rawvideo osd fractions osd level it didn t happen with the previous libass release
0
32,908
8,970,618,733
IssuesEvent
2019-01-29 14:06:01
jscert/jsexplain
https://api.github.com/repos/jscert/jsexplain
closed
`make init` fails when coming from ocaml 4.06
bug buildsystem
Biniou fails to build, but I think this is a problem with jbuilder: ``` [ERROR] The compilation of biniou failed at "jbuilder build -p biniou -j 1". #=== ERROR while installing biniou.1.2.0 ======================================# # opam-version 1.2.2 # os darwin # command jbuilder build -p biniou -j 1 # path /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0 # compiler 4.04.2 # exit-code 1 # env-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.env # stdout-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.out # stderr-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.err ### stderr ### # [...] # File "src/bi_io.ml", line 1: # Error: /Users/schmitta/.opam/4.06.1/lib/easy-format/easy_format.cmi # is not a compiled interface for this version of OCaml. # It seems to be for a newer version of OCaml. # ocamlc src/.biniou.objs/bi_io.{cmo,cmt} (exit 2) # (cd _build/default && /Users/schmitta/.opam/4.04.2/bin/ocamlc.opt -safe-string -g -bin-annot -I src/.biniou.objs -I /Users/schmitta/.opam/4.06.1/lib/easy-format -no-alias-deps -o src/.biniou.objs/bi_io.cmo -c -impl src/bi_io.ml) # File "src/bi_io.ml", line 1: # Error: /Users/schmitta/.opam/4.06.1/lib/easy-format/easy_format.cmi # is not a compiled interface for this version of OCaml. # It seems to be for a newer version of OCaml. ```
1.0
`make init` fails when coming from ocaml 4.06 - Biniou fails to build, but I think this is a problem with jbuilder: ``` [ERROR] The compilation of biniou failed at "jbuilder build -p biniou -j 1". #=== ERROR while installing biniou.1.2.0 ======================================# # opam-version 1.2.2 # os darwin # command jbuilder build -p biniou -j 1 # path /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0 # compiler 4.04.2 # exit-code 1 # env-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.env # stdout-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.out # stderr-file /Users/schmitta/.opam/4.04.2/build/biniou.1.2.0/biniou-95433-a71bb4.err ### stderr ### # [...] # File "src/bi_io.ml", line 1: # Error: /Users/schmitta/.opam/4.06.1/lib/easy-format/easy_format.cmi # is not a compiled interface for this version of OCaml. # It seems to be for a newer version of OCaml. # ocamlc src/.biniou.objs/bi_io.{cmo,cmt} (exit 2) # (cd _build/default && /Users/schmitta/.opam/4.04.2/bin/ocamlc.opt -safe-string -g -bin-annot -I src/.biniou.objs -I /Users/schmitta/.opam/4.06.1/lib/easy-format -no-alias-deps -o src/.biniou.objs/bi_io.cmo -c -impl src/bi_io.ml) # File "src/bi_io.ml", line 1: # Error: /Users/schmitta/.opam/4.06.1/lib/easy-format/easy_format.cmi # is not a compiled interface for this version of OCaml. # It seems to be for a newer version of OCaml. ```
non_priority
make init fails when coming from ocaml biniou fails to build but i think this is a problem with jbuilder the compilation of biniou failed at jbuilder build p biniou j error while installing biniou opam version os darwin command jbuilder build p biniou j path users schmitta opam build biniou compiler exit code env file users schmitta opam build biniou biniou env stdout file users schmitta opam build biniou biniou out stderr file users schmitta opam build biniou biniou err stderr file src bi io ml line error users schmitta opam lib easy format easy format cmi is not a compiled interface for this version of ocaml it seems to be for a newer version of ocaml ocamlc src biniou objs bi io cmo cmt exit cd build default users schmitta opam bin ocamlc opt safe string g bin annot i src biniou objs i users schmitta opam lib easy format no alias deps o src biniou objs bi io cmo c impl src bi io ml file src bi io ml line error users schmitta opam lib easy format easy format cmi is not a compiled interface for this version of ocaml it seems to be for a newer version of ocaml
0
93,569
15,895,256,956
IssuesEvent
2021-04-11 13:22:45
rammatzkvosky/caseflow
https://api.github.com/repos/rammatzkvosky/caseflow
opened
CVE-2020-28481 (Medium) detected in socket.io-1.7.3.tgz
security vulnerability
## CVE-2020-28481 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-1.7.3.tgz</b></p></summary> <p>node.js realtime framework server</p> <p>Library home page: <a href="https://registry.npmjs.org/socket.io/-/socket.io-1.7.3.tgz">https://registry.npmjs.org/socket.io/-/socket.io-1.7.3.tgz</a></p> <p>Path to dependency file: caseflow/client/package.json</p> <p>Path to vulnerable library: caseflow/client/node_modules/socket.io/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - :x: **socket.io-1.7.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/caseflow/commit/8f9f47fb6b371540f9f6b5421908ffe6200fad09">8f9f47fb6b371540f9f6b5421908ffe6200fad09</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default. <p>Publish Date: 2021-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481>CVE-2020-28481</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481</a></p> <p>Release Date: 2021-01-19</p> <p>Fix Resolution: 2.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"socket.io","packageVersion":"1.7.3","packageFilePaths":["/client/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.7.1;socket.io:1.7.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28481","vulnerabilityDetails":"The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481","cvss3Severity":"medium","cvss3Score":"4.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-28481 (Medium) detected in socket.io-1.7.3.tgz - ## CVE-2020-28481 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-1.7.3.tgz</b></p></summary> <p>node.js realtime framework server</p> <p>Library home page: <a href="https://registry.npmjs.org/socket.io/-/socket.io-1.7.3.tgz">https://registry.npmjs.org/socket.io/-/socket.io-1.7.3.tgz</a></p> <p>Path to dependency file: caseflow/client/package.json</p> <p>Path to vulnerable library: caseflow/client/node_modules/socket.io/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - :x: **socket.io-1.7.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/caseflow/commit/8f9f47fb6b371540f9f6b5421908ffe6200fad09">8f9f47fb6b371540f9f6b5421908ffe6200fad09</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default. <p>Publish Date: 2021-01-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481>CVE-2020-28481</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28481</a></p> <p>Release Date: 2021-01-19</p> <p>Fix Resolution: 2.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"socket.io","packageVersion":"1.7.3","packageFilePaths":["/client/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma:1.7.1;socket.io:1.7.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28481","vulnerabilityDetails":"The package socket.io before 2.4.0 are vulnerable to Insecure Defaults due to CORS Misconfiguration. All domains are whitelisted by default.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28481","cvss3Severity":"medium","cvss3Score":"4.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in socket io tgz cve medium severity vulnerability vulnerable library socket io tgz node js realtime framework server library home page a href path to dependency file caseflow client package json path to vulnerable library caseflow client node modules socket io package json dependency hierarchy karma tgz root library x socket io tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package socket io before are vulnerable to insecure defaults due to cors misconfiguration all domains are whitelisted by default publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree karma socket io isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the package socket io before are vulnerable to insecure defaults due to cors misconfiguration all domains are whitelisted by default vulnerabilityurl
0
198,767
14,996,668,728
IssuesEvent
2021-01-29 15:53:23
civictechindex/CTI-website-frontend
https://api.github.com/repos/civictechindex/CTI-website-frontend
closed
Clean up code in ESLint & pass Code Climate
bug role: front end role: testing
### Overview We need to clean up code in ESLint & Code Climate so it can clear compilation issues. ### Action Items - [x] Review action steps with Jonathan/ Laurence - [ ] Clean up code in ESLint - [ ] Confirm it can pass Code Climate ### Resources/Instructions Useful command for auto-fixing via ESLint (low hanging formatting fruit): "lint": "eslint 'src/**/*.js'", "lint:fix": "eslint --fix 'src/**/*.js'",
1.0
Clean up code in ESLint & pass Code Climate - ### Overview We need to clean up code in ESLint & Code Climate so it can clear compilation issues. ### Action Items - [x] Review action steps with Jonathan/ Laurence - [ ] Clean up code in ESLint - [ ] Confirm it can pass Code Climate ### Resources/Instructions Useful command for auto-fixing via ESLint (low hanging formatting fruit): "lint": "eslint 'src/**/*.js'", "lint:fix": "eslint --fix 'src/**/*.js'",
non_priority
clean up code in eslint pass code climate overview we need to clean up code in eslint code climate so it can clear compilation issues action items review action steps with jonathan laurence clean up code in eslint confirm it can pass code climate resources instructions useful command for auto fixing via eslint low hanging formatting fruit lint eslint src js lint fix eslint fix src js
0
11,812
5,093,985,361
IssuesEvent
2017-01-03 09:33:06
pymor/pymor
https://api.github.com/repos/pymor/pymor
closed
pyevtk install
buildsystem
We still depend on pyevtk being installed like `pip3 install https://bitbucket.org/pauloh/pyevtk/get/tip.tar.gz` instead of just `pip3 install pyevtk`. This is not really reflected anywhere. And it's also unclear when/if there will be a new pypi release of it.
1.0
pyevtk install - We still depend on pyevtk being installed like `pip3 install https://bitbucket.org/pauloh/pyevtk/get/tip.tar.gz` instead of just `pip3 install pyevtk`. This is not really reflected anywhere. And it's also unclear when/if there will be a new pypi release of it.
non_priority
pyevtk install we still depend on pyevtk being installed like install instead of just install pyevtk this is not really reflected anywhere and it s also unclear when if there will be a new pypi release of it
0
108,775
16,822,669,870
IssuesEvent
2021-06-17 14:45:29
idonthaveafifaaddiction/flink
https://api.github.com/repos/idonthaveafifaaddiction/flink
opened
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz
security vulnerability
## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: flink/flink-runtime-web/web-dashboard/package.json</p> <p>Path to vulnerable library: flink/flink-runtime-web/web-dashboard/node_modules/base/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/micromatch/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/nanomatch/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/define-property/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/extglob/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/clone-deep/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/snapdragon-node/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - lint-staged-8.1.5.tgz (Root Library) - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","packageFilePaths":["/flink-runtime-web/web-dashboard/package.json"],"isTransitiveDependency":true,"dependencyTree":"lint-staged:8.1.5;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-20149 (High) detected in kind-of-6.0.2.tgz - ## CVE-2019-20149 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>kind-of-6.0.2.tgz</b></p></summary> <p>Get the native type of a value.</p> <p>Library home page: <a href="https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz">https://registry.npmjs.org/kind-of/-/kind-of-6.0.2.tgz</a></p> <p>Path to dependency file: flink/flink-runtime-web/web-dashboard/package.json</p> <p>Path to vulnerable library: flink/flink-runtime-web/web-dashboard/node_modules/base/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/micromatch/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/nanomatch/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/define-property/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/extglob/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/clone-deep/node_modules/kind-of/package.json,flink/flink-runtime-web/web-dashboard/node_modules/snapdragon-node/node_modules/kind-of/package.json</p> <p> Dependency Hierarchy: - lint-staged-8.1.5.tgz (Root Library) - micromatch-3.1.10.tgz - :x: **kind-of-6.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by 'constructor': {'name':'Symbol'}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result. <p>Publish Date: 2019-12-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149>CVE-2019-20149</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-20149</a></p> <p>Release Date: 2019-12-30</p> <p>Fix Resolution: 6.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"kind-of","packageVersion":"6.0.2","packageFilePaths":["/flink-runtime-web/web-dashboard/package.json"],"isTransitiveDependency":true,"dependencyTree":"lint-staged:8.1.5;micromatch:3.1.10;kind-of:6.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-20149","vulnerabilityDetails":"ctorName in index.js in kind-of v6.0.2 allows external user input to overwrite certain internal attributes via a conflicting name, as demonstrated by \u0027constructor\u0027: {\u0027name\u0027:\u0027Symbol\u0027}. Hence, a crafted payload can overwrite this builtin attribute to manipulate the type detection result.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20149","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in kind of tgz cve high severity vulnerability vulnerable library kind of tgz get the native type of a value library home page a href path to dependency file flink flink runtime web web dashboard package json path to vulnerable library flink flink runtime web web dashboard node modules base node modules kind of package json flink flink runtime web web dashboard node modules micromatch node modules kind of package json flink flink runtime web web dashboard node modules nanomatch node modules kind of package json flink flink runtime web web dashboard node modules define property node modules kind of package json flink flink runtime web web dashboard node modules extglob node modules kind of package json flink flink runtime web web dashboard node modules clone deep node modules kind of package json flink flink runtime web web dashboard node modules snapdragon node node modules kind of package json dependency hierarchy lint staged tgz root library micromatch tgz x kind of tgz vulnerable library found in head commit a href found in base branch master vulnerability details ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by constructor name symbol hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree lint staged micromatch kind of isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails ctorname in index js in kind of allows external user input to overwrite certain internal attributes via a conflicting name as demonstrated by hence a crafted payload can overwrite this builtin attribute to manipulate the type detection result vulnerabilityurl
0
99,294
8,696,246,464
IssuesEvent
2018-12-04 16:59:10
phetsims/beers-law-lab
https://api.github.com/repos/phetsims/beers-law-lab
closed
CT: iframe api failed
type:automated-testing
``` beers-law-lab : phet-io-tests : assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543913814604:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543913814604:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930172303:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930172303:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543910575918:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543910575918:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543924312729:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543924312729:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930986454:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930986454:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM ```
1.0
CT: iframe api failed - ``` beers-law-lab : phet-io-tests : assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543913814604:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543913814604:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930172303:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930172303:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543910575918:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543910575918:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543924312729:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543924312729:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM beers-law-lab : phet-io-tests : no-assert 6 out of 7 tests passed. 1 failed. SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.concentrationScreen.view.soluteControls.soluteComboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930986454:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) SimTests: beers-law-lab: iframe api failed: phetioDocumentation expected for beersLawLab.beersLawScreen.view.solutionControls.comboBox at https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/js/tests/SimTests.js?bust=1543930986454:269:20 at dispatch (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:304:56) at windowMessageListener (https://bayes.colorado.edu/continuous-testing/snapshot-1543903744057/phet-io-wrappers/common/js/Client.js:364:13) id: Bayes Chrome Approximately 12/3/2018, 11:09:04 PM ```
non_priority
ct iframe api failed beers law lab phet io tests assert out of tests passed failed simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab concentrationscreen view solutecontrols solutecombobox at at dispatch at windowmessagelistener simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab beerslawscreen view solutioncontrols combobox at at dispatch at windowmessagelistener id bayes chrome approximately pm beers law lab phet io tests assert out of tests passed failed simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab concentrationscreen view solutecontrols solutecombobox at at dispatch at windowmessagelistener simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab beerslawscreen view solutioncontrols combobox at at dispatch at windowmessagelistener id bayes chrome approximately pm beers law lab phet io tests no assert out of tests passed failed simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab concentrationscreen view solutecontrols solutecombobox at at dispatch at windowmessagelistener simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab beerslawscreen view solutioncontrols combobox at at dispatch at windowmessagelistener id bayes chrome approximately pm beers law lab phet io tests no assert out of tests passed failed simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab concentrationscreen view solutecontrols solutecombobox at at dispatch at windowmessagelistener simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab beerslawscreen view solutioncontrols combobox at at dispatch at windowmessagelistener id bayes chrome approximately pm beers law lab phet io tests no assert out of tests passed failed simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab concentrationscreen view solutecontrols solutecombobox at at dispatch at windowmessagelistener simtests beers law lab iframe api failed phetiodocumentation expected for beerslawlab beerslawscreen view solutioncontrols combobox at at dispatch at windowmessagelistener id bayes chrome approximately pm
0
339,077
24,609,690,499
IssuesEvent
2022-10-14 19:57:17
simonw/datasette
https://api.github.com/repos/simonw/datasette
closed
Update screenshots in documentation to match latest designs
documentation
https://docs.datasette.io/en/0.62/full_text_search.html has this out-of-date screenshot: <img width="1191" alt="image" src="https://user-images.githubusercontent.com/9599/195911747-386f4cd2-5239-4c83-8e0c-072e6ae56ff6.png">
1.0
Update screenshots in documentation to match latest designs - https://docs.datasette.io/en/0.62/full_text_search.html has this out-of-date screenshot: <img width="1191" alt="image" src="https://user-images.githubusercontent.com/9599/195911747-386f4cd2-5239-4c83-8e0c-072e6ae56ff6.png">
non_priority
update screenshots in documentation to match latest designs has this out of date screenshot img width alt image src
0
36,252
17,576,319,794
IssuesEvent
2021-08-15 17:25:07
surge-synthesizer/surge
https://api.github.com/repos/surge-synthesizer/surge
closed
Surge allocates 3 oscillators for each voice
Performance
The bug fix in 973741c0885254c8025e118dd1b365f9528817db stops oscillators from leaking for each voice played by reseting the std::uniq_ptr at the freeing of over voice. But why were they being allocated at all? That's a problem we should fix. Here's how voice allocation works in surge 1. At the outset surge allocates enough memory for MAX_VOICES voices and puts them in a list called voices_array (`std::array<std::array<SurgeVoice,MAX_VOICES>,2>`). It tracks voices in use per scene with `voices_used_by` 2. When a voice is taken off the shelf Surge does an in-place constructor call (look for the `new (nvoice)` in SurgeSynthesizer) which doesn't allocate memory but just runs the constructor. 3. But at some point since that code was written, the oscillators in SurgeVoice became std::uniq_ptr which get allocated with spawn_osc in SurgeVoice. My change in the commit above makes sure they are reset and subsequently freed so we don't loose that memory on each voice 4. But really we should do a similar trick where we recycle oscillators from a pool so we don't have to do 3 allocs on every voice. Although machines are fast enough generally nowadays to tolerate this alloc it's a really bad practice to be doing it on each voice creation when we've worked so hard to not. So should clean this up. Fixing this is hairy. This issue is a more specific version of #505
True
Surge allocates 3 oscillators for each voice - The bug fix in 973741c0885254c8025e118dd1b365f9528817db stops oscillators from leaking for each voice played by reseting the std::uniq_ptr at the freeing of over voice. But why were they being allocated at all? That's a problem we should fix. Here's how voice allocation works in surge 1. At the outset surge allocates enough memory for MAX_VOICES voices and puts them in a list called voices_array (`std::array<std::array<SurgeVoice,MAX_VOICES>,2>`). It tracks voices in use per scene with `voices_used_by` 2. When a voice is taken off the shelf Surge does an in-place constructor call (look for the `new (nvoice)` in SurgeSynthesizer) which doesn't allocate memory but just runs the constructor. 3. But at some point since that code was written, the oscillators in SurgeVoice became std::uniq_ptr which get allocated with spawn_osc in SurgeVoice. My change in the commit above makes sure they are reset and subsequently freed so we don't loose that memory on each voice 4. But really we should do a similar trick where we recycle oscillators from a pool so we don't have to do 3 allocs on every voice. Although machines are fast enough generally nowadays to tolerate this alloc it's a really bad practice to be doing it on each voice creation when we've worked so hard to not. So should clean this up. Fixing this is hairy. This issue is a more specific version of #505
non_priority
surge allocates oscillators for each voice the bug fix in stops oscillators from leaking for each voice played by reseting the std uniq ptr at the freeing of over voice but why were they being allocated at all that s a problem we should fix here s how voice allocation works in surge at the outset surge allocates enough memory for max voices voices and puts them in a list called voices array std array it tracks voices in use per scene with voices used by when a voice is taken off the shelf surge does an in place constructor call look for the new nvoice in surgesynthesizer which doesn t allocate memory but just runs the constructor but at some point since that code was written the oscillators in surgevoice became std uniq ptr which get allocated with spawn osc in surgevoice my change in the commit above makes sure they are reset and subsequently freed so we don t loose that memory on each voice but really we should do a similar trick where we recycle oscillators from a pool so we don t have to do allocs on every voice although machines are fast enough generally nowadays to tolerate this alloc it s a really bad practice to be doing it on each voice creation when we ve worked so hard to not so should clean this up fixing this is hairy this issue is a more specific version of
0
217,796
24,351,596,441
IssuesEvent
2022-10-03 01:00:09
billmcchesney1/smart-test-framework
https://api.github.com/repos/billmcchesney1/smart-test-framework
opened
CVE-2022-42003 (Medium) detected in jackson-databind-2.11.0.jar
security vulnerability
## CVE-2022-42003 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.11.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /stf-misc-lib/pom.xml</p> <p>Path to vulnerable library: /canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.11.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/smart-test-framework/commit/9c749e8542c6194fbfda15186a19f7be23142106">9c749e8542c6194fbfda15186a19f7be23142106</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.14.0-rc1, resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting, when the UNWRAP_SINGLE_VALUE_ARRAYS feature is enabled. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42003>CVE-2022-42003</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-42003 (Medium) detected in jackson-databind-2.11.0.jar - ## CVE-2022-42003 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.11.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /stf-misc-lib/pom.xml</p> <p>Path to vulnerable library: /canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.11.0/jackson-databind-2.11.0.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.11.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/smart-test-framework/commit/9c749e8542c6194fbfda15186a19f7be23142106">9c749e8542c6194fbfda15186a19f7be23142106</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.14.0-rc1, resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting, when the UNWRAP_SINGLE_VALUE_ARRAYS feature is enabled. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42003>CVE-2022-42003</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_priority
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file stf misc lib pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting when the unwrap single value arrays feature is enabled publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
263,330
28,029,968,343
IssuesEvent
2023-03-28 11:44:21
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
reopened
CVE-2020-0431 (Medium) detected in linuxv5.2
Mend: dependency security vulnerability
## CVE-2020-0431 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-input.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-input.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In kbd_keycode of keyboard.c, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android kernelAndroid ID: A-144161459 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0431>CVE-2020-0431</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-0431">https://www.linuxkernelcves.com/cves/CVE-2020-0431</a></p> <p>Release Date: 2020-09-17</p> <p>Fix Resolution: v3.16.83,4.4.210,v4.9.210,v4.14.165,v4.19.96,v5.4.12</p> </p> </details> <p></p>
True
CVE-2020-0431 (Medium) detected in linuxv5.2 - ## CVE-2020-0431 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-input.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-input.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In kbd_keycode of keyboard.c, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android kernelAndroid ID: A-144161459 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0431>CVE-2020-0431</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-0431">https://www.linuxkernelcves.com/cves/CVE-2020-0431</a></p> <p>Release Date: 2020-09-17</p> <p>Fix Resolution: v3.16.83,4.4.210,v4.9.210,v4.14.165,v4.19.96,v5.4.12</p> </p> </details> <p></p>
non_priority
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files drivers hid hid input c drivers hid hid input c vulnerability details in kbd keycode of keyboard c there is a possible out of bounds write due to a missing bounds check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android kernelandroid id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
233,238
25,756,424,137
IssuesEvent
2022-12-08 16:45:14
dotCMS/core
https://api.github.com/repos/dotCMS/core
opened
FE validation needed when trying to add a long value key that also creates a misalignment with the UI
Type : Bug Cat : UI Cat : Security Team: Scout Next Release
### Summary When adding a long key to the value we are getting a db problem which presumably can show too much information from the DB ### Test Data User: [email protected] Password: admin ### Url(s): ### Desktop Browsers: macOS 12.6 FF v106.0.3 ### Preconditions: 1. 2. ### Steps to Reproduce: 1. Go to https://localhost:8443/c 2. 3. 4. ### Expected Result: Adding a FE validation here can solve the issue or a different approach to the endpoint can be a solver also ### Actual Result: We are getting UI issues and a potential security issue ### Reference(s): ![206245718-18f067c3-aecd-4d20-b2b0-337d07b1480d](https://user-images.githubusercontent.com/110418590/206509051-f98311b0-c6bc-42d9-9325-c51ac059e44e.png) ![206246525-00e6c44f-2570-44c9-8f6c-78b57c6dc17a](https://user-images.githubusercontent.com/110418590/206509130-8252b21f-99d2-43f1-955f-1f5dd496b6ae.png) ### Note(s): Quick notes added to the report ### Acceptance criteria 1. Add a FE validation here to fix this error
True
FE validation needed when trying to add a long value key that also creates a misalignment with the UI - ### Summary When adding a long key to the value we are getting a db problem which presumably can show too much information from the DB ### Test Data User: [email protected] Password: admin ### Url(s): ### Desktop Browsers: macOS 12.6 FF v106.0.3 ### Preconditions: 1. 2. ### Steps to Reproduce: 1. Go to https://localhost:8443/c 2. 3. 4. ### Expected Result: Adding a FE validation here can solve the issue or a different approach to the endpoint can be a solver also ### Actual Result: We are getting UI issues and a potential security issue ### Reference(s): ![206245718-18f067c3-aecd-4d20-b2b0-337d07b1480d](https://user-images.githubusercontent.com/110418590/206509051-f98311b0-c6bc-42d9-9325-c51ac059e44e.png) ![206246525-00e6c44f-2570-44c9-8f6c-78b57c6dc17a](https://user-images.githubusercontent.com/110418590/206509130-8252b21f-99d2-43f1-955f-1f5dd496b6ae.png) ### Note(s): Quick notes added to the report ### Acceptance criteria 1. Add a FE validation here to fix this error
non_priority
fe validation needed when trying to add a long value key that also creates a misalignment with the ui summary when adding a long key to the value we are getting a db problem which presumably can show too much information from the db test data user admin dotcms com password admin url s desktop browsers macos ff preconditions steps to reproduce go to expected result adding a fe validation here can solve the issue or a different approach to the endpoint can be a solver also actual result we are getting ui issues and a potential security issue reference s note s quick notes added to the report acceptance criteria add a fe validation here to fix this error
0
11,066
4,877,425,630
IssuesEvent
2016-11-16 15:40:42
CartoDB/cartodb
https://api.github.com/repos/CartoDB/cartodb
closed
Updating prefix/suffix legend's with null fails
bug Builder Legends
### Steps to Reproduce 1. Style a dot map in order to generate a legend 2. Update the legend with a suffix, leaving the prefix blank. ### Current Result An 422 raises. ### Expected result A success request
1.0
Updating prefix/suffix legend's with null fails - ### Steps to Reproduce 1. Style a dot map in order to generate a legend 2. Update the legend with a suffix, leaving the prefix blank. ### Current Result An 422 raises. ### Expected result A success request
non_priority
updating prefix suffix legend s with null fails steps to reproduce style a dot map in order to generate a legend update the legend with a suffix leaving the prefix blank current result an raises expected result a success request
0
105,603
9,087,510,972
IssuesEvent
2019-02-18 13:56:24
edenlabllc/ehealth.api
https://api.github.com/repos/edenlabllc/ehealth.api
closed
New variables for model improving
BE [zube]: In Test kind/improvement project/deduplication status/todo
In order to improve model next steps should be done: - [ ] add to variables `d_document_number_length` and `d_document_distinct` - [ ] change regular expression for documents cleaning - for `BIRTH_CERTIFICATE` `regexp_replace(regexp_replace(lower(pd1.number), '[():*''|",.?<=+\\//\%#№ _-]', '', 'g'),'^[qjJlL1iі!!IІ]','i','g')` - for the rest documents `regexp_replace(lower(pd1.number), '[():*''|",.?<=+\\//\%#№ _-]', '', 'g')` - [ ] change binning
1.0
New variables for model improving - In order to improve model next steps should be done: - [ ] add to variables `d_document_number_length` and `d_document_distinct` - [ ] change regular expression for documents cleaning - for `BIRTH_CERTIFICATE` `regexp_replace(regexp_replace(lower(pd1.number), '[():*''|",.?<=+\\//\%#№ _-]', '', 'g'),'^[qjJlL1iі!!IІ]','i','g')` - for the rest documents `regexp_replace(lower(pd1.number), '[():*''|",.?<=+\\//\%#№ _-]', '', 'g')` - [ ] change binning
non_priority
new variables for model improving in order to improve model next steps should be done add to variables d document number length and d document distinct change regular expression for documents cleaning for birth certificate regexp replace regexp replace lower number g i g for the rest documents regexp replace lower number g change binning
0
248,270
21,007,143,913
IssuesEvent
2022-03-30 00:24:20
facebook/react
https://api.github.com/repos/facebook/react
closed
flushPassiveEffects and IsThisRendererActing are on the wrong "secret" object
Component: Test Utils React Core Team
They're currently attached to the "Events" namespace but they don't have anything to do with events. It needs to move to its own names on the secret object. Once we delete Events, after ReactTestUtils and RNW are fixed, then we won't be able to clean up the array because these two are on it. Additionally the IsSomeRendererActing field is in the production bundles. We need to make all of these DEV-only.
1.0
flushPassiveEffects and IsThisRendererActing are on the wrong "secret" object - They're currently attached to the "Events" namespace but they don't have anything to do with events. It needs to move to its own names on the secret object. Once we delete Events, after ReactTestUtils and RNW are fixed, then we won't be able to clean up the array because these two are on it. Additionally the IsSomeRendererActing field is in the production bundles. We need to make all of these DEV-only.
non_priority
flushpassiveeffects and isthisrendereracting are on the wrong secret object they re currently attached to the events namespace but they don t have anything to do with events it needs to move to its own names on the secret object once we delete events after reacttestutils and rnw are fixed then we won t be able to clean up the array because these two are on it additionally the issomerendereracting field is in the production bundles we need to make all of these dev only
0
12,079
3,574,977,743
IssuesEvent
2016-01-27 14:20:36
geryxyz/CodeMetropolis
https://api.github.com/repos/geryxyz/CodeMetropolis
closed
Expand Installation Guide
documentation
It would be much more easier to install dependencies, if the links were there in case of Java Runtime Environment 8 (http://www.oracle.com/technetwork/java/javase/downloads/jre8-downloads-2133155.html) and Mindecraft client 1.8 (https://minecraft.net/download).
1.0
Expand Installation Guide - It would be much more easier to install dependencies, if the links were there in case of Java Runtime Environment 8 (http://www.oracle.com/technetwork/java/javase/downloads/jre8-downloads-2133155.html) and Mindecraft client 1.8 (https://minecraft.net/download).
non_priority
expand installation guide it would be much more easier to install dependencies if the links were there in case of java runtime environment and mindecraft client
0
346,424
30,913,145,205
IssuesEvent
2023-08-05 01:11:27
CollinHeist/TitleCardMaker-Blueprints
https://api.github.com/repos/CollinHeist/TitleCardMaker-Blueprints
closed
[Blueprint] The Boys
blueprint created passed-tests
### Series Name The Boys ### Series Year 2019 ### Creator Username flowcool ### Blueprint Description StarWars cards with The Boys font. ### Blueprint ```json { "series": { "font_id": 0, "card_type": "star wars", "template_ids": [] }, "episodes": {}, "templates": [], "fonts": [ { "name": "The Boys", "color": "#EC1B17", "delete_missing": true, "file": "Charlie don't surf.ttf", "title_case": "upper" } ], "preview": "preview.jpg" } ``` ### Preview Title Card ![The Boys (2019) - S01E05](https://github.com/CollinHeist/TitleCardMaker-Blueprints/assets/19570361/3d95eefd-7de3-4525-9053-0dad34cad8b3) ### Zip of Font Files [Charlie don't surf.zip](https://github.com/CollinHeist/TitleCardMaker-Blueprints/files/12236647/Charlie.don.t.surf.zip)
1.0
[Blueprint] The Boys - ### Series Name The Boys ### Series Year 2019 ### Creator Username flowcool ### Blueprint Description StarWars cards with The Boys font. ### Blueprint ```json { "series": { "font_id": 0, "card_type": "star wars", "template_ids": [] }, "episodes": {}, "templates": [], "fonts": [ { "name": "The Boys", "color": "#EC1B17", "delete_missing": true, "file": "Charlie don't surf.ttf", "title_case": "upper" } ], "preview": "preview.jpg" } ``` ### Preview Title Card ![The Boys (2019) - S01E05](https://github.com/CollinHeist/TitleCardMaker-Blueprints/assets/19570361/3d95eefd-7de3-4525-9053-0dad34cad8b3) ### Zip of Font Files [Charlie don't surf.zip](https://github.com/CollinHeist/TitleCardMaker-Blueprints/files/12236647/Charlie.don.t.surf.zip)
non_priority
the boys series name the boys series year creator username flowcool blueprint description starwars cards with the boys font blueprint json series font id card type star wars template ids episodes templates fonts name the boys color delete missing true file charlie don t surf ttf title case upper preview preview jpg preview title card zip of font files
0
116,060
11,899,482,623
IssuesEvent
2020-03-30 09:04:46
IjzerenHein/react-navigation-shared-element
https://api.github.com/repos/IjzerenHein/react-navigation-shared-element
closed
[v5] typo in Readme
documentation navigation-5
I suppose `id` should be the same. <img width="658" alt="Screenshot 2020-03-29 at 22 59 20" src="https://user-images.githubusercontent.com/6594232/77859283-1fb22300-7211-11ea-91c7-d4a7c9f64901.png">
1.0
[v5] typo in Readme - I suppose `id` should be the same. <img width="658" alt="Screenshot 2020-03-29 at 22 59 20" src="https://user-images.githubusercontent.com/6594232/77859283-1fb22300-7211-11ea-91c7-d4a7c9f64901.png">
non_priority
typo in readme i suppose id should be the same img width alt screenshot at src
0
114,113
14,531,200,849
IssuesEvent
2020-12-14 20:25:01
jupyterlab/jupyterlab
https://api.github.com/repos/jupyterlab/jupyterlab
closed
Feature request: an option to change the font
status:Needs Info tag:Design and UX type:Bug
**Describe the feature** I know that you can change the font and its size if you go to `Settings -> Advanced Settings Editor -> Notebook`. And there are three options: * Code Cell Configuration * Markdown Cell Configuration * Raw Cell Configuration But please, look at this screenshot (it looks small, please open it up in another tab) ![different_fonts](https://user-images.githubusercontent.com/33338133/48987130-03b4e380-f0ea-11e8-980f-659912cc9344.png) The font of the output cell is different from the input cell. For example, one of the cells has the code `model_1.check_test_point()` which shows a font (changed by me) and the output of that cell is ``` mu_interval__ -1.39 sigma_interval__ -1.39 y_bar -1350.85 Name: Log-probability of test_point, dtype: float64 ``` which has another font. Maybe you can't see it, but I can assure you that the fonts are different. The goal of this feature (if it is implemented) is... well, to have the option to change that font. **Desktop (please complete the following information):** - OS: Ubuntu 16.04.5 LTS - Browser: Firefox 63.0.3 - JupyterLab: 0.35.4
1.0
Feature request: an option to change the font - **Describe the feature** I know that you can change the font and its size if you go to `Settings -> Advanced Settings Editor -> Notebook`. And there are three options: * Code Cell Configuration * Markdown Cell Configuration * Raw Cell Configuration But please, look at this screenshot (it looks small, please open it up in another tab) ![different_fonts](https://user-images.githubusercontent.com/33338133/48987130-03b4e380-f0ea-11e8-980f-659912cc9344.png) The font of the output cell is different from the input cell. For example, one of the cells has the code `model_1.check_test_point()` which shows a font (changed by me) and the output of that cell is ``` mu_interval__ -1.39 sigma_interval__ -1.39 y_bar -1350.85 Name: Log-probability of test_point, dtype: float64 ``` which has another font. Maybe you can't see it, but I can assure you that the fonts are different. The goal of this feature (if it is implemented) is... well, to have the option to change that font. **Desktop (please complete the following information):** - OS: Ubuntu 16.04.5 LTS - Browser: Firefox 63.0.3 - JupyterLab: 0.35.4
non_priority
feature request an option to change the font describe the feature i know that you can change the font and its size if you go to settings advanced settings editor notebook and there are three options code cell configuration markdown cell configuration raw cell configuration but please look at this screenshot it looks small please open it up in another tab the font of the output cell is different from the input cell for example one of the cells has the code model check test point which shows a font changed by me and the output of that cell is mu interval sigma interval y bar name log probability of test point dtype which has another font maybe you can t see it but i can assure you that the fonts are different the goal of this feature if it is implemented is well to have the option to change that font desktop please complete the following information os ubuntu lts browser firefox jupyterlab
0
390,094
26,847,343,391
IssuesEvent
2023-02-03 08:18:37
Carlos-SE/App-chores
https://api.github.com/repos/Carlos-SE/App-chores
closed
Test runner
documentation enhancement objetivo-4
Es necesario elegir un `test runner` que nos permita crear y ejecutar pruebas a nuestro código, de forma que se tenga en cuenta los principio F.I.R.S.T (Fast, independent, repeatable, self-validating, timely). Documentar las diferentes opciones que se planteen y la elección de la que se usará.
1.0
Test runner - Es necesario elegir un `test runner` que nos permita crear y ejecutar pruebas a nuestro código, de forma que se tenga en cuenta los principio F.I.R.S.T (Fast, independent, repeatable, self-validating, timely). Documentar las diferentes opciones que se planteen y la elección de la que se usará.
non_priority
test runner es necesario elegir un test runner que nos permita crear y ejecutar pruebas a nuestro código de forma que se tenga en cuenta los principio f i r s t fast independent repeatable self validating timely documentar las diferentes opciones que se planteen y la elección de la que se usará
0
136,543
11,049,536,762
IssuesEvent
2019-12-10 00:03:53
MangopearUK/European-Boating-Association--Theme
https://api.github.com/repos/MangopearUK/European-Boating-Association--Theme
closed
Test & audit: 2016-GA2-Vienna
Testing
Page URL: https://eba.eu.com/membership/meetings/2016-ga2-vienna/ ## Table of contents - [x] **Task 1:** Perform automated audits _(10 tasks)_ - [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_ - [x] **Task 3:** Breakpoint testing _(15 tasks)_ - [x] **Task 4:** Re-run automated audits _(10 tasks)_ ## 1: Perform automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors ## 2: Manual standards & accessibility tests _(61 tasks)_ ### Forms - [x] Give all form elements permanently visible labels - [x] Place labels above form elements - [x] Mark invalid fields clearly and provide associated error messages - [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode - [x] Ensure all form fields have the correct requried state - [x] Provide status and error messages as WAI-ARIA live regions ### Readability of content - [x] Ensure page has good grammar - [x] Ensure page content has been spell-checked - [x] Make sure headings are in logical order - [x] Ensure the same content is available across different devices and platforms - [x] Begin long, multi-section documents with a table of contents ### Presentation - [x] Make sure all content is formatted correctly - [x] Avoid all-caps text - [x] Make sure data tables wider than their container can be scrolled horizontally - [x] Use the same design patterns to solve the same problems - [x] Do not mark up subheadings/straplines with separate heading elements ### Links & buttons #### Links - [x] Check all links to ensure they work - [x] Check all links to third party websites use `rel="noopener"` - [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility" - [x] Provide a skip link if necessary - [x] Underline links — at least in body copy - [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label) #### Buttons - [x] Ensure primary calls to action are easy to recognize and reach - [x] Provide clear, unambiguous focus styles - [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software - [x] Ensure disabled controls are not focusable - [x] Make sure controls within hidden content are not focusable - [x] Provide large touch "targets" for interactive elements - [x] Make controls look like controls; give them strong perceived affordance - [x] Use well-established, therefore recognizable, icons and symbols ### Assistive technology - [x] Ensure content is not obscured through zooming - [x] Support Windows high contrast mode (use images, not background images) - [x] Provide alternative text for salient images - [x] Make scrollable elements focusable for keyboard users - [x] Ensure keyboard focus order is logical regarding visual layout - [x] Match semantics to behavior for assistive technology users - [x] Provide a default language and use lang="[ISO code]" for subsections in different languages - [x] Inform the user when there are important changes to the application state - [x] Do not hijack standard scrolling behavior - [x] Do not instate "infinite scroll" by default; provide buttons to load more items ### General accessibility - [x] Make sure text and background colors contrast sufficiently - [x] Do not rely on color for differentiation of visual elements - [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech - [x] Provide a print stylesheet - [x] Honour requests to remove animation via the prefers-reduced-motion media query ### SEO - [x] Ensure all pages have appropriate title - [x] Ensure all pages have meta descriptions - [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype) - [x] Check whether page should be appearing in sitemap - [x] Make sure page has Facebook and Twitter large image previews set correctly - [x] Check canonical links for page - [x] Mark as cornerstone content? ### Performance - [x] Ensure all CSS assets are minified and concatenated - [x] Ensure all JS assets are minified and concatenated - [x] Ensure all images are compressed - [x] Where possible, remove redundant code - [x] Ensure all SVG assets have been optimised - [x] Make sure styles and scripts are not render blocking - [x] Ensure large image assets are lazy loaded ### Other - [x] Make sure all content belongs to a landmark element - [x] Provide a manifest.json file for identifiable homescreen entries ## 3: Breakpoint testing _(15 tasks)_ ### Desktop - [x] Provide a full screenshot of **1920px** wide page - [x] Provide a full screenshot of **1500px** wide page - [x] Provide a full screenshot of **1280px** wide page - [x] Provide a full screenshot of **1024px** wide page ### Tablet - [x] Provide a full screenshot of **960px** wide page - [x] Provide a full screenshot of **800px** wide page - [x] Provide a full screenshot of **760px** wide page - [x] Provide a full screenshot of **650px** wide page ### Mobile - [x] Provide a full screenshot of **600px** wide page - [x] Provide a full screenshot of **500px** wide page - [x] Provide a full screenshot of **450px** wide page - [x] Provide a full screenshot of **380px** wide page - [x] Provide a full screenshot of **320px** wide page - [x] Provide a full screenshot of **280px** wide page - [x] Provide a full screenshot of **250px** wide page ## 4: Re-run automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors
1.0
Test & audit: 2016-GA2-Vienna - Page URL: https://eba.eu.com/membership/meetings/2016-ga2-vienna/ ## Table of contents - [x] **Task 1:** Perform automated audits _(10 tasks)_ - [x] **Task 2:** Manual standards & accessibility tests _(61 tasks)_ - [x] **Task 3:** Breakpoint testing _(15 tasks)_ - [x] **Task 4:** Re-run automated audits _(10 tasks)_ ## 1: Perform automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors ## 2: Manual standards & accessibility tests _(61 tasks)_ ### Forms - [x] Give all form elements permanently visible labels - [x] Place labels above form elements - [x] Mark invalid fields clearly and provide associated error messages - [x] Make forms as short as possible; offer shortcuts like autocompleting the address using the postcode - [x] Ensure all form fields have the correct requried state - [x] Provide status and error messages as WAI-ARIA live regions ### Readability of content - [x] Ensure page has good grammar - [x] Ensure page content has been spell-checked - [x] Make sure headings are in logical order - [x] Ensure the same content is available across different devices and platforms - [x] Begin long, multi-section documents with a table of contents ### Presentation - [x] Make sure all content is formatted correctly - [x] Avoid all-caps text - [x] Make sure data tables wider than their container can be scrolled horizontally - [x] Use the same design patterns to solve the same problems - [x] Do not mark up subheadings/straplines with separate heading elements ### Links & buttons #### Links - [x] Check all links to ensure they work - [x] Check all links to third party websites use `rel="noopener"` - [x] Make sure the purpose of a link is clearly described: "read more" vs. "read more about accessibility" - [x] Provide a skip link if necessary - [x] Underline links — at least in body copy - [x] Warn users of links that have unusual behaviors, like linking off-site, or loading a new tab (i.e. aria-label) #### Buttons - [x] Ensure primary calls to action are easy to recognize and reach - [x] Provide clear, unambiguous focus styles - [x] Ensure states (pressed, expanded, invalid, etc) are communicated to assistive software - [x] Ensure disabled controls are not focusable - [x] Make sure controls within hidden content are not focusable - [x] Provide large touch "targets" for interactive elements - [x] Make controls look like controls; give them strong perceived affordance - [x] Use well-established, therefore recognizable, icons and symbols ### Assistive technology - [x] Ensure content is not obscured through zooming - [x] Support Windows high contrast mode (use images, not background images) - [x] Provide alternative text for salient images - [x] Make scrollable elements focusable for keyboard users - [x] Ensure keyboard focus order is logical regarding visual layout - [x] Match semantics to behavior for assistive technology users - [x] Provide a default language and use lang="[ISO code]" for subsections in different languages - [x] Inform the user when there are important changes to the application state - [x] Do not hijack standard scrolling behavior - [x] Do not instate "infinite scroll" by default; provide buttons to load more items ### General accessibility - [x] Make sure text and background colors contrast sufficiently - [x] Do not rely on color for differentiation of visual elements - [x] Avoid images of text — text that cannot be translated, selected, or understood by assistive tech - [x] Provide a print stylesheet - [x] Honour requests to remove animation via the prefers-reduced-motion media query ### SEO - [x] Ensure all pages have appropriate title - [x] Ensure all pages have meta descriptions - [x] Make content easier to find and improve search results with structured data [Read more](https://developers.google.com/search/docs/guides/prototype) - [x] Check whether page should be appearing in sitemap - [x] Make sure page has Facebook and Twitter large image previews set correctly - [x] Check canonical links for page - [x] Mark as cornerstone content? ### Performance - [x] Ensure all CSS assets are minified and concatenated - [x] Ensure all JS assets are minified and concatenated - [x] Ensure all images are compressed - [x] Where possible, remove redundant code - [x] Ensure all SVG assets have been optimised - [x] Make sure styles and scripts are not render blocking - [x] Ensure large image assets are lazy loaded ### Other - [x] Make sure all content belongs to a landmark element - [x] Provide a manifest.json file for identifiable homescreen entries ## 3: Breakpoint testing _(15 tasks)_ ### Desktop - [x] Provide a full screenshot of **1920px** wide page - [x] Provide a full screenshot of **1500px** wide page - [x] Provide a full screenshot of **1280px** wide page - [x] Provide a full screenshot of **1024px** wide page ### Tablet - [x] Provide a full screenshot of **960px** wide page - [x] Provide a full screenshot of **800px** wide page - [x] Provide a full screenshot of **760px** wide page - [x] Provide a full screenshot of **650px** wide page ### Mobile - [x] Provide a full screenshot of **600px** wide page - [x] Provide a full screenshot of **500px** wide page - [x] Provide a full screenshot of **450px** wide page - [x] Provide a full screenshot of **380px** wide page - [x] Provide a full screenshot of **320px** wide page - [x] Provide a full screenshot of **280px** wide page - [x] Provide a full screenshot of **250px** wide page ## 4: Re-run automated audits _(10 tasks)_ ### Lighthouse: - [x] Run "Accessibility" audit in lighthouse _(using incognito tab)_ - [x] Run "Performance" audit in lighthouse _(using incognito tab)_ - [x] Run "Best practices" audit in lighthouse _(using incognito tab)_ - [x] Run "SEO" audit in lighthouse _(using incognito tab)_ - [x] Run "PWA" audit in lighthouse _(using incognito tab)_ ### Pingdom - [x] Run full audit of the the page's performance in Pingdom ### Browser's console - [x] Check Chrome's console for errors ### Log results of audits - [x] Screenshot snapshot of the lighthouse audits - [x] Upload PDF of detailed lighthouse reports - [x] Provide a screenshot of any console errors
non_priority
test audit vienna page url table of contents task perform automated audits tasks task manual standards accessibility tests tasks task breakpoint testing tasks task re run automated audits tasks perform automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors manual standards accessibility tests tasks forms give all form elements permanently visible labels place labels above form elements mark invalid fields clearly and provide associated error messages make forms as short as possible offer shortcuts like autocompleting the address using the postcode ensure all form fields have the correct requried state provide status and error messages as wai aria live regions readability of content ensure page has good grammar ensure page content has been spell checked make sure headings are in logical order ensure the same content is available across different devices and platforms begin long multi section documents with a table of contents presentation make sure all content is formatted correctly avoid all caps text make sure data tables wider than their container can be scrolled horizontally use the same design patterns to solve the same problems do not mark up subheadings straplines with separate heading elements links buttons links check all links to ensure they work check all links to third party websites use rel noopener make sure the purpose of a link is clearly described read more vs read more about accessibility provide a skip link if necessary underline links — at least in body copy warn users of links that have unusual behaviors like linking off site or loading a new tab i e aria label buttons ensure primary calls to action are easy to recognize and reach provide clear unambiguous focus styles ensure states pressed expanded invalid etc are communicated to assistive software ensure disabled controls are not focusable make sure controls within hidden content are not focusable provide large touch targets for interactive elements make controls look like controls give them strong perceived affordance use well established therefore recognizable icons and symbols assistive technology ensure content is not obscured through zooming support windows high contrast mode use images not background images provide alternative text for salient images make scrollable elements focusable for keyboard users ensure keyboard focus order is logical regarding visual layout match semantics to behavior for assistive technology users provide a default language and use lang for subsections in different languages inform the user when there are important changes to the application state do not hijack standard scrolling behavior do not instate infinite scroll by default provide buttons to load more items general accessibility make sure text and background colors contrast sufficiently do not rely on color for differentiation of visual elements avoid images of text — text that cannot be translated selected or understood by assistive tech provide a print stylesheet honour requests to remove animation via the prefers reduced motion media query seo ensure all pages have appropriate title ensure all pages have meta descriptions make content easier to find and improve search results with structured data check whether page should be appearing in sitemap make sure page has facebook and twitter large image previews set correctly check canonical links for page mark as cornerstone content performance ensure all css assets are minified and concatenated ensure all js assets are minified and concatenated ensure all images are compressed where possible remove redundant code ensure all svg assets have been optimised make sure styles and scripts are not render blocking ensure large image assets are lazy loaded other make sure all content belongs to a landmark element provide a manifest json file for identifiable homescreen entries breakpoint testing tasks desktop provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page tablet provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page mobile provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page provide a full screenshot of wide page re run automated audits tasks lighthouse run accessibility audit in lighthouse using incognito tab run performance audit in lighthouse using incognito tab run best practices audit in lighthouse using incognito tab run seo audit in lighthouse using incognito tab run pwa audit in lighthouse using incognito tab pingdom run full audit of the the page s performance in pingdom browser s console check chrome s console for errors log results of audits screenshot snapshot of the lighthouse audits upload pdf of detailed lighthouse reports provide a screenshot of any console errors
0
103,443
11,356,553,712
IssuesEvent
2020-01-24 23:08:46
commfish/seak_sablefish
https://api.github.com/repos/commfish/seak_sablefish
closed
odd trip_no in 2019 data
documentation
Sent to M Vaughn 2019-12-19: 1) the A/B trip numbers for 113 and 144. I don’t have any records of A/B trips for other years. This will break a lot of code… Is there a good reason to have these separate? If so, why not assign them separate trip nos? 2) Trip 136 with 6 lb? Seems odd.
1.0
odd trip_no in 2019 data - Sent to M Vaughn 2019-12-19: 1) the A/B trip numbers for 113 and 144. I don’t have any records of A/B trips for other years. This will break a lot of code… Is there a good reason to have these separate? If so, why not assign them separate trip nos? 2) Trip 136 with 6 lb? Seems odd.
non_priority
odd trip no in data sent to m vaughn the a b trip numbers for and i don’t have any records of a b trips for other years this will break a lot of code… is there a good reason to have these separate if so why not assign them separate trip nos trip with lb seems odd
0
213,169
16,504,113,245
IssuesEvent
2021-05-25 17:06:33
DataBiosphere/hca-metadata-api
https://api.github.com/repos/DataBiosphere/hca-metadata-api
closed
Add ability to use canned staging areas for unit tests
dcp demo demoed enh no demo orange test
@danielsotirhos to discuss details with @hannes-ucsc
1.0
Add ability to use canned staging areas for unit tests - @danielsotirhos to discuss details with @hannes-ucsc
non_priority
add ability to use canned staging areas for unit tests danielsotirhos to discuss details with hannes ucsc
0
178,905
21,509,581,160
IssuesEvent
2022-04-28 01:56:53
eldorplus/conception-website
https://api.github.com/repos/eldorplus/conception-website
closed
CVE-2019-16769 (Medium) detected in serialize-javascript-1.7.0.tgz - autoclosed
security vulnerability
## CVE-2019-16769 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.7.0.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz</a></p> <p>Path to dependency file: /conception-website/package.json</p> <p>Path to vulnerable library: conception-website/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - copy-webpack-plugin-4.6.0.tgz (Root Library) - :x: **serialize-javascript-1.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The serialize-javascript npm package before version 2.1.1 is vulnerable to Cross-site Scripting (XSS). It does not properly mitigate against unsafe characters in serialized regular expressions. This vulnerability is not affected on Node.js environment since Node.js's implementation of RegExp.prototype.toString() backslash-escapes all forward slashes in regular expressions. If serialized data of regular expression objects are used in an environment other than Node.js, it is affected by this vulnerability. <p>Publish Date: 2019-12-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16769>CVE-2019-16769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: v2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-16769 (Medium) detected in serialize-javascript-1.7.0.tgz - autoclosed - ## CVE-2019-16769 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.7.0.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.7.0.tgz</a></p> <p>Path to dependency file: /conception-website/package.json</p> <p>Path to vulnerable library: conception-website/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - copy-webpack-plugin-4.6.0.tgz (Root Library) - :x: **serialize-javascript-1.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/eldorplus/conception-website/commit/d01989566d2a70e542b08bc943c5de9d223ce39d">d01989566d2a70e542b08bc943c5de9d223ce39d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The serialize-javascript npm package before version 2.1.1 is vulnerable to Cross-site Scripting (XSS). It does not properly mitigate against unsafe characters in serialized regular expressions. This vulnerability is not affected on Node.js environment since Node.js's implementation of RegExp.prototype.toString() backslash-escapes all forward slashes in regular expressions. If serialized data of regular expression objects are used in an environment other than Node.js, it is affected by this vulnerability. <p>Publish Date: 2019-12-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16769>CVE-2019-16769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: v2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in serialize javascript tgz autoclosed cve medium severity vulnerability vulnerable library serialize javascript tgz serialize javascript to a superset of json that includes regular expressions and functions library home page a href path to dependency file conception website package json path to vulnerable library conception website node modules serialize javascript package json dependency hierarchy copy webpack plugin tgz root library x serialize javascript tgz vulnerable library found in head commit a href vulnerability details the serialize javascript npm package before version is vulnerable to cross site scripting xss it does not properly mitigate against unsafe characters in serialized regular expressions this vulnerability is not affected on node js environment since node js s implementation of regexp prototype tostring backslash escapes all forward slashes in regular expressions if serialized data of regular expression objects are used in an environment other than node js it is affected by this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
315
5,313,712,687
IssuesEvent
2017-02-13 13:05:35
DevExpress/testcafe
https://api.github.com/repos/DevExpress/testcafe
opened
Scrolling issues in iOS simulator
AREA: client BROWSER: iOS/Safari SYSTEM: automations
### Are you requesting a feature or reporting a bug? Bug ### What is the current behavior? Due to terrible performance of Safari inside iOS Simulator (Saucelabs is most affected by this), target elements sometimes can't get precise scrolling positions, especially on large pages. ### What is the expected behavior? Scrolling performed precisely on iOS Simulator. ### How would you reproduce the current behavior (if this is a bug)? Perform several actions with scrolling on a large page in iOS simulator on Saucelabs. ### Specify your * operating system: iOS 8, iOS 9 * testcafe version: 0.13.0-alpha2 * node.js version: 7.5.0
1.0
Scrolling issues in iOS simulator - ### Are you requesting a feature or reporting a bug? Bug ### What is the current behavior? Due to terrible performance of Safari inside iOS Simulator (Saucelabs is most affected by this), target elements sometimes can't get precise scrolling positions, especially on large pages. ### What is the expected behavior? Scrolling performed precisely on iOS Simulator. ### How would you reproduce the current behavior (if this is a bug)? Perform several actions with scrolling on a large page in iOS simulator on Saucelabs. ### Specify your * operating system: iOS 8, iOS 9 * testcafe version: 0.13.0-alpha2 * node.js version: 7.5.0
non_priority
scrolling issues in ios simulator are you requesting a feature or reporting a bug bug what is the current behavior due to terrible performance of safari inside ios simulator saucelabs is most affected by this target elements sometimes can t get precise scrolling positions especially on large pages what is the expected behavior scrolling performed precisely on ios simulator how would you reproduce the current behavior if this is a bug perform several actions with scrolling on a large page in ios simulator on saucelabs specify your operating system ios ios testcafe version node js version
0
32,481
13,854,713,147
IssuesEvent
2020-10-15 09:53:34
ITISFoundation/osparc-simcore
https://api.github.com/repos/ITISFoundation/osparc-simcore
opened
Release jupyters to deployed systems
a:pipeline-services
After the recent changes to the jupyters you are responsible for 2 out of 3. For the following services: - jupiter-ocatave-python-math with current published version 1.4.1 (on dalco) and 1.4.0 in aws staging and production (check this link) - jupyter-smash with current published version of 2.2.4 (version is only for dalco) please check here I would propose a release to the fowling versions an all the target platforms: - jupiter-ocatave-python-math release version 1.4.3 (available on master for testing) - jupyter-smash release version 2.2.5 (available on master for testing) You would need to do the following: - check that the new services function as expected and especially that their nodeports are working correctly - I will step in and sync them to the target release platforms - you will then give the access rights to all the people who need to use these services on all the platforms Hope this was helpful! Regards,
1.0
Release jupyters to deployed systems - After the recent changes to the jupyters you are responsible for 2 out of 3. For the following services: - jupiter-ocatave-python-math with current published version 1.4.1 (on dalco) and 1.4.0 in aws staging and production (check this link) - jupyter-smash with current published version of 2.2.4 (version is only for dalco) please check here I would propose a release to the fowling versions an all the target platforms: - jupiter-ocatave-python-math release version 1.4.3 (available on master for testing) - jupyter-smash release version 2.2.5 (available on master for testing) You would need to do the following: - check that the new services function as expected and especially that their nodeports are working correctly - I will step in and sync them to the target release platforms - you will then give the access rights to all the people who need to use these services on all the platforms Hope this was helpful! Regards,
non_priority
release jupyters to deployed systems after the recent changes to the jupyters you are responsible for out of for the following services jupiter ocatave python math with current published version on dalco and in aws staging and production check this link jupyter smash with current published version of version is only for dalco please check here i would propose a release to the fowling versions an all the target platforms jupiter ocatave python math release version available on master for testing jupyter smash release version available on master for testing you would need to do the following check that the new services function as expected and especially that their nodeports are working correctly i will step in and sync them to the target release platforms you will then give the access rights to all the people who need to use these services on all the platforms hope this was helpful regards
0
325,833
24,062,968,557
IssuesEvent
2022-09-17 04:17:08
openkfw/TruBudget
https://api.github.com/repos/openkfw/TruBudget
closed
Documentation mistakes in TB-website
bug wontfix documentation blockchain
- [x] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate. ## Description 😯 The documentation section ["Connect to an existing Blockchain network using docker"](https://trubudget.net/docs/operation-administration/installation/connect-to-an-existing-network/docker) has some mistakes. The descriptions for EXTERNAL_IP and P2P_HOST are not correct, and should be fixed as follows: <img width="1534" alt="image" src="https://user-images.githubusercontent.com/48566702/178290001-8ee4c585-4d0d-4838-a76b-91c41fc76b4d.png"> - EXTERNAL_IP: this environement variable should be set on the **Alpha** blockchain node and should contain the public ip address and port of the **Alpha** blockchain. - P2P_HOST: this environment variable should be set on the **Beta** blockchain node and should contain the public IP-address of the **Alpha** node (where the node should connect to) Also, the descriptions should be more clear to avoid misinterpretation ## Your Environment 🌎 | Tech | Version | | --------------- | ------- | | TruBudget | v2.0.0 |
1.0
Documentation mistakes in TB-website - - [x] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate. ## Description 😯 The documentation section ["Connect to an existing Blockchain network using docker"](https://trubudget.net/docs/operation-administration/installation/connect-to-an-existing-network/docker) has some mistakes. The descriptions for EXTERNAL_IP and P2P_HOST are not correct, and should be fixed as follows: <img width="1534" alt="image" src="https://user-images.githubusercontent.com/48566702/178290001-8ee4c585-4d0d-4838-a76b-91c41fc76b4d.png"> - EXTERNAL_IP: this environement variable should be set on the **Alpha** blockchain node and should contain the public ip address and port of the **Alpha** blockchain. - P2P_HOST: this environment variable should be set on the **Beta** blockchain node and should contain the public IP-address of the **Alpha** node (where the node should connect to) Also, the descriptions should be more clear to avoid misinterpretation ## Your Environment 🌎 | Tech | Version | | --------------- | ------- | | TruBudget | v2.0.0 |
non_priority
documentation mistakes in tb website i have searched the of this repository and believe that this is not a duplicate description 😯 the documentation section has some mistakes the descriptions for external ip and host are not correct and should be fixed as follows img width alt image src external ip this environement variable should be set on the alpha blockchain node and should contain the public ip address and port of the alpha blockchain host this environment variable should be set on the beta blockchain node and should contain the public ip address of the alpha node where the node should connect to also the descriptions should be more clear to avoid misinterpretation your environment 🌎 tech version trubudget
0
99,652
12,463,088,580
IssuesEvent
2020-05-28 10:00:02
ajency/Finaegis-Backend
https://api.github.com/repos/ajency/Finaegis-Backend
closed
Profile Information - No spacing is there between the Notification Bar and the text 'Profile Information'
Android Application Assigned to QA Design Issue Must do Version - 0.0.6
**Describe the Issue** No spacing is there between the Notification Bar and the text 'Profile Information' **Screenshots** ![image](https://user-images.githubusercontent.com/52652632/83104556-a3d03c00-a0d6-11ea-95a3-8cf9941a75e6.png)
1.0
Profile Information - No spacing is there between the Notification Bar and the text 'Profile Information' - **Describe the Issue** No spacing is there between the Notification Bar and the text 'Profile Information' **Screenshots** ![image](https://user-images.githubusercontent.com/52652632/83104556-a3d03c00-a0d6-11ea-95a3-8cf9941a75e6.png)
non_priority
profile information no spacing is there between the notification bar and the text profile information describe the issue no spacing is there between the notification bar and the text profile information screenshots
0
305,829
23,132,789,515
IssuesEvent
2022-07-28 11:58:06
AgnostiqHQ/covalent
https://api.github.com/repos/AgnostiqHQ/covalent
closed
MNIST tutorial fails to train correctly when using Covalent
documentation
As others have already pointed out in [a different issue](https://github.com/AgnostiqHQ/covalent/issues/638). The MNIST tutorial fails to train the ML model when using Covalent. [Cell 14](https://covalent.readthedocs.io/en/latest/tutorials/machine_learning/mnist_classifier.html#Transform-and-run-workflow-with-Covalent) shows a training accuracy of approximately 10%, the same accuracy as from random guessing. The reason is that we're "using Covalent wrong". Users must remember that electrons cannot manipulate external data structures passed by reference. Since the electron can potentially run in a separate process or even on a separate computer from the Covalent server, any references to objects returned from another electron (which are stored in the Covalent server's memory) will surely break. Consequently, **Covalent must make deep copies of all inputs to each electron***. Here's an excerpt from **Cell 10** of the tutorial: ``` optimizer = get_optimizer(model=model, learning_rate=learning_rate, momentum=momentum) model, optimizer = train_model( train_dataloader=train_dataloader, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, model=model, optimizer=optimizer, loss_fn=loss_fn, epochs=epochs, ) ``` `get_optimizer` and `train_model` are separate electrons, defined in Cells 6 and 9 respectively. Here's `get_optimizer`: ``` def get_optimizer( model: NeuralNetwork, learning_rate: float, momentum: float ) -> torch.optim.Optimizer: """Get Stochastic Gradient Descent optimizer.""" return torch.optim.SGD(model.parameters(), learning_rate, momentum) ``` Now let's take a look at `train_model`: ``` def train_model( train_dataloader: torch.utils.data.dataloader.DataLoader, train_losses: list, train_counter: int, log_interval: int, model: NeuralNetwork, loss_fn: callable, epochs: int, results_dir: str = "~/data/mnist/results/", ) -> Tuple[NeuralNetwork,]: ... for epoch in range(1, epochs + 1): print(f"Epoch {epoch}\n-------------------------------") model, optimizer = train_over_one_epoch( dataloader=train_dataloader, model=model, optimizer=optimizer, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, epoch=epoch, loss_fn=loss_fn, ) ... ``` When this electron is executed, the `optimizer` argument refers to a **different** Optimizer instance from what was returned by `get_optimizer` in Cell 10. Indeed if we want to run this electron using the Slurm executor or in a separate worker process using Dask, we must make deep copies of all inputs; a remote server has no way to manipulate the Optimizer instance living on the Covalent server. Similarly, `model` is an independent copy of the `model` defined in the original workflow. **These two objects are not connected**. When `train_model` calls `optimizer.step()` (see `train_over_one_epoch`), the `optimizer` object modifies its own copy of the model parameters while the parameters in `model` remain unchanged. One remedy is to eliminate the `get_optimizer` electron and instead instantiate the Optimizer inside `train_model` using the model passed to `train_model`, something like ``` @ct.electron def train_model( train_dataloader: torch.utils.data.dataloader.DataLoader, train_losses: list, train_counter: int, log_interval: int, model: NeuralNetwork, learning_rate: float, momentum: float, optimizer: torch.optim.Optimizer, loss_fn: callable, epochs: int, results_dir: str = "~/data/mnist/results/", ) -> Tuple[NeuralNetwork,]: """Train neural network model.""" optimizer = torch.optim.SGD(model.parameters(), learning_rate, momentum) for epoch in range(1, epochs + 1): print(f"Epoch {epoch}\n-------------------------------") model, optimizer = train_over_one_epoch( dataloader=train_dataloader, model=model, optimizer=optimizer, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, epoch=epoch, loss_fn=loss_fn, ) ... return model, optimizer ``` *This tutorial worked in older versions of Covalent (v32.3 and before) because the local executor that was used by default didn't make deep copies of inputs to electrons. I would expect the tutorial to fail if one had tried a remote executor for `train_model`. [This commit](https://github.com/AgnostiqHQ/covalent/commit/1a4fa1e99d186629cc439fbf30792ce7a9ffa801#diff-a99ca6e887d5a6439efcda540e9e2661cdafa9e3a92109f2ac9c390299f794fbR197) introduced (effectively) a deep copy for all executors, including local. And nowadays, we default to a multiprocessing-based Dask executor which would require copies of inputs even without that commit.
1.0
MNIST tutorial fails to train correctly when using Covalent - As others have already pointed out in [a different issue](https://github.com/AgnostiqHQ/covalent/issues/638). The MNIST tutorial fails to train the ML model when using Covalent. [Cell 14](https://covalent.readthedocs.io/en/latest/tutorials/machine_learning/mnist_classifier.html#Transform-and-run-workflow-with-Covalent) shows a training accuracy of approximately 10%, the same accuracy as from random guessing. The reason is that we're "using Covalent wrong". Users must remember that electrons cannot manipulate external data structures passed by reference. Since the electron can potentially run in a separate process or even on a separate computer from the Covalent server, any references to objects returned from another electron (which are stored in the Covalent server's memory) will surely break. Consequently, **Covalent must make deep copies of all inputs to each electron***. Here's an excerpt from **Cell 10** of the tutorial: ``` optimizer = get_optimizer(model=model, learning_rate=learning_rate, momentum=momentum) model, optimizer = train_model( train_dataloader=train_dataloader, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, model=model, optimizer=optimizer, loss_fn=loss_fn, epochs=epochs, ) ``` `get_optimizer` and `train_model` are separate electrons, defined in Cells 6 and 9 respectively. Here's `get_optimizer`: ``` def get_optimizer( model: NeuralNetwork, learning_rate: float, momentum: float ) -> torch.optim.Optimizer: """Get Stochastic Gradient Descent optimizer.""" return torch.optim.SGD(model.parameters(), learning_rate, momentum) ``` Now let's take a look at `train_model`: ``` def train_model( train_dataloader: torch.utils.data.dataloader.DataLoader, train_losses: list, train_counter: int, log_interval: int, model: NeuralNetwork, loss_fn: callable, epochs: int, results_dir: str = "~/data/mnist/results/", ) -> Tuple[NeuralNetwork,]: ... for epoch in range(1, epochs + 1): print(f"Epoch {epoch}\n-------------------------------") model, optimizer = train_over_one_epoch( dataloader=train_dataloader, model=model, optimizer=optimizer, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, epoch=epoch, loss_fn=loss_fn, ) ... ``` When this electron is executed, the `optimizer` argument refers to a **different** Optimizer instance from what was returned by `get_optimizer` in Cell 10. Indeed if we want to run this electron using the Slurm executor or in a separate worker process using Dask, we must make deep copies of all inputs; a remote server has no way to manipulate the Optimizer instance living on the Covalent server. Similarly, `model` is an independent copy of the `model` defined in the original workflow. **These two objects are not connected**. When `train_model` calls `optimizer.step()` (see `train_over_one_epoch`), the `optimizer` object modifies its own copy of the model parameters while the parameters in `model` remain unchanged. One remedy is to eliminate the `get_optimizer` electron and instead instantiate the Optimizer inside `train_model` using the model passed to `train_model`, something like ``` @ct.electron def train_model( train_dataloader: torch.utils.data.dataloader.DataLoader, train_losses: list, train_counter: int, log_interval: int, model: NeuralNetwork, learning_rate: float, momentum: float, optimizer: torch.optim.Optimizer, loss_fn: callable, epochs: int, results_dir: str = "~/data/mnist/results/", ) -> Tuple[NeuralNetwork,]: """Train neural network model.""" optimizer = torch.optim.SGD(model.parameters(), learning_rate, momentum) for epoch in range(1, epochs + 1): print(f"Epoch {epoch}\n-------------------------------") model, optimizer = train_over_one_epoch( dataloader=train_dataloader, model=model, optimizer=optimizer, train_losses=train_losses, train_counter=train_counter, log_interval=log_interval, epoch=epoch, loss_fn=loss_fn, ) ... return model, optimizer ``` *This tutorial worked in older versions of Covalent (v32.3 and before) because the local executor that was used by default didn't make deep copies of inputs to electrons. I would expect the tutorial to fail if one had tried a remote executor for `train_model`. [This commit](https://github.com/AgnostiqHQ/covalent/commit/1a4fa1e99d186629cc439fbf30792ce7a9ffa801#diff-a99ca6e887d5a6439efcda540e9e2661cdafa9e3a92109f2ac9c390299f794fbR197) introduced (effectively) a deep copy for all executors, including local. And nowadays, we default to a multiprocessing-based Dask executor which would require copies of inputs even without that commit.
non_priority
mnist tutorial fails to train correctly when using covalent as others have already pointed out in the mnist tutorial fails to train the ml model when using covalent shows a training accuracy of approximately the same accuracy as from random guessing the reason is that we re using covalent wrong users must remember that electrons cannot manipulate external data structures passed by reference since the electron can potentially run in a separate process or even on a separate computer from the covalent server any references to objects returned from another electron which are stored in the covalent server s memory will surely break consequently covalent must make deep copies of all inputs to each electron here s an excerpt from cell of the tutorial optimizer get optimizer model model learning rate learning rate momentum momentum model optimizer train model train dataloader train dataloader train losses train losses train counter train counter log interval log interval model model optimizer optimizer loss fn loss fn epochs epochs get optimizer and train model are separate electrons defined in cells and respectively here s get optimizer def get optimizer model neuralnetwork learning rate float momentum float torch optim optimizer get stochastic gradient descent optimizer return torch optim sgd model parameters learning rate momentum now let s take a look at train model def train model train dataloader torch utils data dataloader dataloader train losses list train counter int log interval int model neuralnetwork loss fn callable epochs int results dir str data mnist results tuple for epoch in range epochs print f epoch epoch n model optimizer train over one epoch dataloader train dataloader model model optimizer optimizer train losses train losses train counter train counter log interval log interval epoch epoch loss fn loss fn when this electron is executed the optimizer argument refers to a different optimizer instance from what was returned by get optimizer in cell indeed if we want to run this electron using the slurm executor or in a separate worker process using dask we must make deep copies of all inputs a remote server has no way to manipulate the optimizer instance living on the covalent server similarly model is an independent copy of the model defined in the original workflow these two objects are not connected when train model calls optimizer step see train over one epoch the optimizer object modifies its own copy of the model parameters while the parameters in model remain unchanged one remedy is to eliminate the get optimizer electron and instead instantiate the optimizer inside train model using the model passed to train model something like ct electron def train model train dataloader torch utils data dataloader dataloader train losses list train counter int log interval int model neuralnetwork learning rate float momentum float optimizer torch optim optimizer loss fn callable epochs int results dir str data mnist results tuple train neural network model optimizer torch optim sgd model parameters learning rate momentum for epoch in range epochs print f epoch epoch n model optimizer train over one epoch dataloader train dataloader model model optimizer optimizer train losses train losses train counter train counter log interval log interval epoch epoch loss fn loss fn return model optimizer this tutorial worked in older versions of covalent and before because the local executor that was used by default didn t make deep copies of inputs to electrons i would expect the tutorial to fail if one had tried a remote executor for train model introduced effectively a deep copy for all executors including local and nowadays we default to a multiprocessing based dask executor which would require copies of inputs even without that commit
0
262,045
27,850,878,690
IssuesEvent
2023-03-20 18:35:42
jgeraigery/kconnect
https://api.github.com/repos/jgeraigery/kconnect
opened
CVE-2020-8565 (Medium) detected in k8s.io/client-go-v0.19.1
Mend: dependency security vulnerability
## CVE-2020-8565 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>k8s.io/client-go-v0.19.1</b></p></summary> <p>Go client for Kubernetes.</p> <p>Library home page: <a href="https://proxy.golang.org/k8s.io/client-go/@v/v0.19.1.zip">https://proxy.golang.org/k8s.io/client-go/@v/v0.19.1.zip</a></p> <p> Dependency Hierarchy: - :x: **k8s.io/client-go-v0.19.1** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/kconnect/commit/28f03af9f116211eb5fc096657fda435b32a3386">28f03af9f116211eb5fc096657fda435b32a3386</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Kubernetes, if the logging level is set to at least 9, authorization and bearer tokens will be written to log files. This can occur both in API server logs and client tool output like kubectl. This affects <= v1.19.3, <= v1.18.10, <= v1.17.13, < v1.20.0-alpha2. <p>Publish Date: 2020-12-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8565>CVE-2020-8565</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0064">https://osv.dev/vulnerability/GO-2020-0064</a></p> <p>Release Date: 2020-12-07</p> <p>Fix Resolution: v1.20.0-alpha.2</p> </p> </details> <p></p>
True
CVE-2020-8565 (Medium) detected in k8s.io/client-go-v0.19.1 - ## CVE-2020-8565 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>k8s.io/client-go-v0.19.1</b></p></summary> <p>Go client for Kubernetes.</p> <p>Library home page: <a href="https://proxy.golang.org/k8s.io/client-go/@v/v0.19.1.zip">https://proxy.golang.org/k8s.io/client-go/@v/v0.19.1.zip</a></p> <p> Dependency Hierarchy: - :x: **k8s.io/client-go-v0.19.1** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/kconnect/commit/28f03af9f116211eb5fc096657fda435b32a3386">28f03af9f116211eb5fc096657fda435b32a3386</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Kubernetes, if the logging level is set to at least 9, authorization and bearer tokens will be written to log files. This can occur both in API server logs and client tool output like kubectl. This affects <= v1.19.3, <= v1.18.10, <= v1.17.13, < v1.20.0-alpha2. <p>Publish Date: 2020-12-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8565>CVE-2020-8565</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GO-2020-0064">https://osv.dev/vulnerability/GO-2020-0064</a></p> <p>Release Date: 2020-12-07</p> <p>Fix Resolution: v1.20.0-alpha.2</p> </p> </details> <p></p>
non_priority
cve medium detected in io client go cve medium severity vulnerability vulnerable library io client go go client for kubernetes library home page a href dependency hierarchy x io client go vulnerable library found in head commit a href found in base branch main vulnerability details in kubernetes if the logging level is set to at least authorization and bearer tokens will be written to log files this can occur both in api server logs and client tool output like kubectl this affects publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution alpha
0
92,407
11,642,379,920
IssuesEvent
2020-02-29 06:53:14
broccolijs/broccoli
https://api.github.com/repos/broccolijs/broccoli
closed
Plugin API a Bit Confusing
enhancement needs feature design plugin
> By convention, plugins will export a function that takes one or more input trees, and returns an output tree object. Usually your plugin will be implemented as a class representing a tree, but it is recommended to make the new operator optional (example). I believe Broccoli walks the tree using regular function calls. Which means that if you have any kind of state being managed in the plugin class that it's not "recommended to make the new operator optional" but is actually required. Does that sound right? I'm finding the API for plugins a bit unusual. If others feel the same then maybe we could come up with something better. What's unusual for me is that the plugin api is just a function while typically being implemented as a class. I personally like using class constructs to manage state but then the optional new operator that is practically mandatory makes it weird. I know that condition is occasionally used in libs to prevent errors if the consumer forgets to use `new`. One example of where this could be a problem is with ES6 classes. You can't invoke an ES6 class even though it's just syntactic sugar for a regular constructor function. You could just use functions and capture state with a closure, which I'm not really a fan of. You could even go as far as not allowing state and just using pure functions but I doubt that's feasible and I think that the state enables a nice and intuitive API. It also seems like the plugin object and the tree object are a hybrid (at least with the cache-writer plugin), is that correct? Why not just define a plugin object interface and a tree object interface and allow any plugin creators to implement those interfaces as objects however they want to? Am I just completely missing the point? I'm not an idiot (or at least not completely :smile:) so I stand by it being a bit confusing and so at least could maybe use some more documentation.
1.0
Plugin API a Bit Confusing - > By convention, plugins will export a function that takes one or more input trees, and returns an output tree object. Usually your plugin will be implemented as a class representing a tree, but it is recommended to make the new operator optional (example). I believe Broccoli walks the tree using regular function calls. Which means that if you have any kind of state being managed in the plugin class that it's not "recommended to make the new operator optional" but is actually required. Does that sound right? I'm finding the API for plugins a bit unusual. If others feel the same then maybe we could come up with something better. What's unusual for me is that the plugin api is just a function while typically being implemented as a class. I personally like using class constructs to manage state but then the optional new operator that is practically mandatory makes it weird. I know that condition is occasionally used in libs to prevent errors if the consumer forgets to use `new`. One example of where this could be a problem is with ES6 classes. You can't invoke an ES6 class even though it's just syntactic sugar for a regular constructor function. You could just use functions and capture state with a closure, which I'm not really a fan of. You could even go as far as not allowing state and just using pure functions but I doubt that's feasible and I think that the state enables a nice and intuitive API. It also seems like the plugin object and the tree object are a hybrid (at least with the cache-writer plugin), is that correct? Why not just define a plugin object interface and a tree object interface and allow any plugin creators to implement those interfaces as objects however they want to? Am I just completely missing the point? I'm not an idiot (or at least not completely :smile:) so I stand by it being a bit confusing and so at least could maybe use some more documentation.
non_priority
plugin api a bit confusing by convention plugins will export a function that takes one or more input trees and returns an output tree object usually your plugin will be implemented as a class representing a tree but it is recommended to make the new operator optional example i believe broccoli walks the tree using regular function calls which means that if you have any kind of state being managed in the plugin class that it s not recommended to make the new operator optional but is actually required does that sound right i m finding the api for plugins a bit unusual if others feel the same then maybe we could come up with something better what s unusual for me is that the plugin api is just a function while typically being implemented as a class i personally like using class constructs to manage state but then the optional new operator that is practically mandatory makes it weird i know that condition is occasionally used in libs to prevent errors if the consumer forgets to use new one example of where this could be a problem is with classes you can t invoke an class even though it s just syntactic sugar for a regular constructor function you could just use functions and capture state with a closure which i m not really a fan of you could even go as far as not allowing state and just using pure functions but i doubt that s feasible and i think that the state enables a nice and intuitive api it also seems like the plugin object and the tree object are a hybrid at least with the cache writer plugin is that correct why not just define a plugin object interface and a tree object interface and allow any plugin creators to implement those interfaces as objects however they want to am i just completely missing the point i m not an idiot or at least not completely smile so i stand by it being a bit confusing and so at least could maybe use some more documentation
0
130,560
10,617,607,794
IssuesEvent
2019-10-12 20:20:47
Vachok/ftpplus
https://api.github.com/repos/Vachok/ftpplus
closed
testGetPossibleVariantsOfPC [D254]
Lowest TestQuality bug mint resolution_Fixed resolution_Wont Do
Execute LocalUserResolverTest::testGetPossibleVariantsOfPC**testGetPossibleVariantsOfPC** *LocalUserResolverTest* * did not expect to find [true] but found [false]* *java.lang.AssertionError*
1.0
testGetPossibleVariantsOfPC [D254] - Execute LocalUserResolverTest::testGetPossibleVariantsOfPC**testGetPossibleVariantsOfPC** *LocalUserResolverTest* * did not expect to find [true] but found [false]* *java.lang.AssertionError*
non_priority
testgetpossiblevariantsofpc execute localuserresolvertest testgetpossiblevariantsofpc testgetpossiblevariantsofpc localuserresolvertest did not expect to find but found java lang assertionerror
0
152,710
13,465,335,293
IssuesEvent
2020-09-09 20:41:39
cookpad/terraform-aws-eks
https://api.github.com/repos/cookpad/terraform-aws-eks
closed
Document / script a procedure to service out asg_node_group
documentation enhancement
Something like: * Disable cluster autoscaler for asg_node_group #111 * Discover all the nodes managed by the asg_node_group (perhaps we need to add a label for this) * Cordon all the nodes * Drain each node with a configurable delay between * Remove the asg_node_group module!
1.0
Document / script a procedure to service out asg_node_group - Something like: * Disable cluster autoscaler for asg_node_group #111 * Discover all the nodes managed by the asg_node_group (perhaps we need to add a label for this) * Cordon all the nodes * Drain each node with a configurable delay between * Remove the asg_node_group module!
non_priority
document script a procedure to service out asg node group something like disable cluster autoscaler for asg node group discover all the nodes managed by the asg node group perhaps we need to add a label for this cordon all the nodes drain each node with a configurable delay between remove the asg node group module
0
40,592
12,799,580,712
IssuesEvent
2020-07-02 15:36:59
TreyM-WSS/WhiteSource-Demo
https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo
opened
CVE-2016-9878 (High) detected in spring-webmvc-4.3.2.RELEASE.jar
security vulnerability
## CVE-2016-9878 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-webmvc-4.3.2.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/4.3.2.RELEASE/spring-webmvc-4.3.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **spring-webmvc-4.3.2.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Pivotal Spring Framework before 3.2.18, 4.2.x before 4.2.9, and 4.3.x before 4.3.5. Paths provided to the ResourceServlet were not properly sanitized and as a result exposed to directory traversal attacks. <p>Publish Date: 2016-12-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9878>CVE-2016-9878</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9878">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9878</a></p> <p>Release Date: 2016-12-29</p> <p>Fix Resolution: 3.2.18,4.2.9,4.3.5.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-webmvc","packageVersion":"4.3.2.RELEASE","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework:spring-webmvc:4.3.2.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.18,4.2.9,4.3.5."}],"vulnerabilityIdentifier":"CVE-2016-9878","vulnerabilityDetails":"An issue was discovered in Pivotal Spring Framework before 3.2.18, 4.2.x before 4.2.9, and 4.3.x before 4.3.5. Paths provided to the ResourceServlet were not properly sanitized and as a result exposed to directory traversal attacks.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9878","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-9878 (High) detected in spring-webmvc-4.3.2.RELEASE.jar - ## CVE-2016-9878 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-webmvc-4.3.2.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/WhiteSource-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/4.3.2.RELEASE/spring-webmvc-4.3.2.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **spring-webmvc-4.3.2.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TreyM-WSS/WhiteSource-Demo/commits/75659f691fb82d67ecd666ba6076394defeb92d0">75659f691fb82d67ecd666ba6076394defeb92d0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Pivotal Spring Framework before 3.2.18, 4.2.x before 4.2.9, and 4.3.x before 4.3.5. Paths provided to the ResourceServlet were not properly sanitized and as a result exposed to directory traversal attacks. <p>Publish Date: 2016-12-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9878>CVE-2016-9878</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9878">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9878</a></p> <p>Release Date: 2016-12-29</p> <p>Fix Resolution: 3.2.18,4.2.9,4.3.5.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-webmvc","packageVersion":"4.3.2.RELEASE","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;org.springframework:spring-webmvc:4.3.2.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.2.18,4.2.9,4.3.5."}],"vulnerabilityIdentifier":"CVE-2016-9878","vulnerabilityDetails":"An issue was discovered in Pivotal Spring Framework before 3.2.18, 4.2.x before 4.2.9, and 4.3.x before 4.3.5. Paths provided to the ResourceServlet were not properly sanitized and as a result exposed to directory traversal attacks.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9878","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in spring webmvc release jar cve high severity vulnerability vulnerable library spring webmvc release jar spring web mvc library home page a href path to dependency file tmp ws scm whitesource demo pom xml path to vulnerable library home wss scanner repository org springframework spring webmvc release spring webmvc release jar dependency hierarchy spring boot starter web release jar root library x spring webmvc release jar vulnerable library found in head commit a href vulnerability details an issue was discovered in pivotal spring framework before x before and x before paths provided to the resourceservlet were not properly sanitized and as a result exposed to directory traversal attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in pivotal spring framework before x before and x before paths provided to the resourceservlet were not properly sanitized and as a result exposed to directory traversal attacks vulnerabilityurl
0
15,409
19,598,429,494
IssuesEvent
2022-01-05 21:01:03
goodboy/tractor
https://api.github.com/repos/goodboy/tractor
opened
Deprecate `ActorNursery.run_in_actor()` and offer as part of a *wrapper* cluster API
help wanted testing api supervision cancellation process_spawning
I've rambled on wayyy too long in #287 but thanks to attempts to resolve that issue *I think* this is the way I'd like to go on a so called *one shot task per actor*-style "worker pool" API. #### TDLR premises - `Portal.result()` + `ActorNursery.run_in_actor()` couples a *future*-like API into our nursery's actor spawning interface at the *wrong level of abstraction* - `.run_in_actor()` can be implemented as a syntactic sugar on top of `ActorNursery.start_actor()` + `Portal.run()` using `trio` tasks in a similar way to our [`concurrent.futures` worker pool example](https://github.com/goodboy/tractor/blob/master/examples/parallelism/concurrent_actors_primes.py#L108) - conducting [error collection and propagation in our `ActorNursery`'s teardown machinery](https://github.com/goodboy/tractor/blob/master/tractor/_supervise.py#L431) is entirely superfluous and a *path dependent* legacy design which has a few pretty awful side effects: - it makes the nursery need [to be aware of `.run_in_actor()` portals](https://github.com/goodboy/tractor/blob/master/tractor/_supervise.py#L217) - it enforces duplication between `.run_in_actor()` and `Portal.run()` - it complicates [spawn machinery with special cases](https://github.com/goodboy/tractor/blob/master/tractor/_spawn.py#L348) - it results in [hard to maintain tests]() due to indeterminacy in cancelled vs. errored child results (which btw won't go away if we change the API layering but at least we can leverage std `trio` nursery machinery instead of rolling our own 😂) #### ToDO - [ ] write up a version of `.run_in_actor()` style on top of the rest of the nursery API - [ ] convert the `nested_multierrors()` test to use the `.open_context()` / `.run()` style and ensure we can still get as reliable of performance - [ ] figure out how to offer the higher level / new `.run_in_actor()` API as a cluster helper
1.0
Deprecate `ActorNursery.run_in_actor()` and offer as part of a *wrapper* cluster API - I've rambled on wayyy too long in #287 but thanks to attempts to resolve that issue *I think* this is the way I'd like to go on a so called *one shot task per actor*-style "worker pool" API. #### TDLR premises - `Portal.result()` + `ActorNursery.run_in_actor()` couples a *future*-like API into our nursery's actor spawning interface at the *wrong level of abstraction* - `.run_in_actor()` can be implemented as a syntactic sugar on top of `ActorNursery.start_actor()` + `Portal.run()` using `trio` tasks in a similar way to our [`concurrent.futures` worker pool example](https://github.com/goodboy/tractor/blob/master/examples/parallelism/concurrent_actors_primes.py#L108) - conducting [error collection and propagation in our `ActorNursery`'s teardown machinery](https://github.com/goodboy/tractor/blob/master/tractor/_supervise.py#L431) is entirely superfluous and a *path dependent* legacy design which has a few pretty awful side effects: - it makes the nursery need [to be aware of `.run_in_actor()` portals](https://github.com/goodboy/tractor/blob/master/tractor/_supervise.py#L217) - it enforces duplication between `.run_in_actor()` and `Portal.run()` - it complicates [spawn machinery with special cases](https://github.com/goodboy/tractor/blob/master/tractor/_spawn.py#L348) - it results in [hard to maintain tests]() due to indeterminacy in cancelled vs. errored child results (which btw won't go away if we change the API layering but at least we can leverage std `trio` nursery machinery instead of rolling our own 😂) #### ToDO - [ ] write up a version of `.run_in_actor()` style on top of the rest of the nursery API - [ ] convert the `nested_multierrors()` test to use the `.open_context()` / `.run()` style and ensure we can still get as reliable of performance - [ ] figure out how to offer the higher level / new `.run_in_actor()` API as a cluster helper
non_priority
deprecate actornursery run in actor and offer as part of a wrapper cluster api i ve rambled on wayyy too long in but thanks to attempts to resolve that issue i think this is the way i d like to go on a so called one shot task per actor style worker pool api tdlr premises portal result actornursery run in actor couples a future like api into our nursery s actor spawning interface at the wrong level of abstraction run in actor can be implemented as a syntactic sugar on top of actornursery start actor portal run using trio tasks in a similar way to our conducting is entirely superfluous and a path dependent legacy design which has a few pretty awful side effects it makes the nursery need it enforces duplication between run in actor and portal run it complicates it results in due to indeterminacy in cancelled vs errored child results which btw won t go away if we change the api layering but at least we can leverage std trio nursery machinery instead of rolling our own 😂 todo write up a version of run in actor style on top of the rest of the nursery api convert the nested multierrors test to use the open context run style and ensure we can still get as reliable of performance figure out how to offer the higher level new run in actor api as a cluster helper
0
267,481
23,303,064,491
IssuesEvent
2022-08-07 16:07:17
JonasMuehlmann/datastructures.go
https://api.github.com/repos/JonasMuehlmann/datastructures.go
closed
Compare lists after operations, to require correctnes
todo tests
Compare lists after operations, to require correctnes https://github.com/JonasMuehlmann/datastructures.go/blob/c5d510817d83387c67548f03bca14f73ab517343/maps/hashbidimap/hashbidimap_test.go#L340 ```golang package hashbidimap import ( "testing" "github.com/JonasMuehlmann/datastructures.go/tests" "github.com/JonasMuehlmann/datastructures.go/utils" "github.com/stretchr/testify/assert" ) func TestRemove(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] newMap *Map[string, int] toRemove string }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), newMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), toRemove: "foo", }, { name: "single item", toRemove: "foo", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item, target does not exist", toRemove: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", toRemove: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "baz": 3}), }, } for _, test := range tests { test.originalMap.Remove(utils.BasicComparator[string], test.toRemove) assert.ElementsMatchf(t, test.originalMap.GetKeys(), test.newMap.GetKeys(), test.name) } } func TestPut(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] newMap *Map[string, int] keyToAdd string valueToAdd int }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), keyToAdd: "foo", valueToAdd: 1, }, { name: "single item", keyToAdd: "foo", valueToAdd: 1, newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item, overwrite", keyToAdd: "foo", valueToAdd: 2, originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 2}), }, { name: "3 items", keyToAdd: "bar", valueToAdd: 2, originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "baz": 3}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { test.originalMap.Put(test.keyToAdd, test.valueToAdd) assert.ElementsMatchf(t, test.originalMap.GetKeys(), test.newMap.GetKeys(), test.name) } } func TestGet(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] keyToGet string value int found bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), keyToGet: "foo", found: false, }, { name: "single item", keyToGet: "foo", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), value: 1, found: true, }, { name: "single item, target does not exist", keyToGet: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), found: false, }, { name: "3 items", keyToGet: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), value: 2, found: true, }, } for _, test := range tests { value, found := test.originalMap.Get(test.keyToGet) assert.Equalf(t, test.value, value, test.name) assert.Equalf(t, test.found, found, test.name) } } func TestGetKeys(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] keys []string }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), keys: []string{}, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), keys: []string{"foo"}, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), keys: []string{"foo", "bar", "baz"}, }, } for _, test := range tests { keys := test.originalMap.GetKeys() assert.ElementsMatchf(t, test.keys, keys, test.name) } } func TestGetValues(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] values []int }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), values: []int{}, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), values: []int{1}, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), values: []int{1, 2, 3}, }, } for _, test := range tests { values := test.originalMap.GetValues() assert.ElementsMatchf(t, test.values, values, test.name) } } func TestIsEmpty(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] isEmpty bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), isEmpty: true, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), isEmpty: false, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), isEmpty: false, }, } for _, test := range tests { isEmpty := test.originalMap.IsEmpty() assert.Equal(t, test.isEmpty, isEmpty, test.name) } } func TestClear(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] isEmptyBefore bool isEmptyAfter bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), isEmptyBefore: true, isEmptyAfter: true, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), isEmptyBefore: false, isEmptyAfter: true, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), isEmptyBefore: false, isEmptyAfter: true, }, } for _, test := range tests { isEmptyBefore := test.originalMap.IsEmpty() assert.Equal(t, test.isEmptyBefore, isEmptyBefore, test.name) test.originalMap.Clear() isEmptAfter := test.originalMap.IsEmpty() assert.Equal(t, test.isEmptyAfter, isEmptAfter, test.name) } } func TestNewFromIterator(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { it := test.originalMap.OrderedBegin(utils.BasicComparator[string]) newMap := NewFromIterator[string, int](utils.BasicComparator[string], utils.BasicComparator[int], it) assert.ElementsMatchf(t, test.originalMap.GetKeys(), newMap.GetKeys(), test.name) } } func TestNewFromIterators(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { first := test.originalMap.OrderedBegin(utils.BasicComparator[string]) end := test.originalMap.OrderedEnd(utils.BasicComparator[string]) newMap := NewFromIterators[string, int](utils.BasicComparator[string], utils.BasicComparator[int], first, end) assert.ElementsMatchf(t, test.originalMap.GetKeys(), newMap.GetKeys(), test.name) } } // TODO: Compare lists after operations, to require correctnes func BenchmarkHashMapRemove(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) for i := 0; i < n; i++ { m.Put(i, "foo") } b.StartTimer() for i := 0; i < n; i++ { m.Remove(utils.BasicComparator[int], i) } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) for i := 0; i < n; i++ { m[i] = "foo" } b.StartTimer() for i := 0; i < n; i++ { delete(m, i) } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } func BenchmarkHashMapGet(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) for i := 0; i < n; i++ { m.Put(i, "foo") } b.StartTimer() for i := 0; i < n; i++ { _, _ = m.Get(i) } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) for i := 0; i < n; i++ { m[i] = "foo" } b.StartTimer() for i := 0; i < n; i++ { _, _ = m[i] } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } func BenchmarkHashMapPut(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) b.StartTimer() for i := 0; i < n; i++ { m.Put(i, "foo") } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) b.StartTimer() for i := 0; i < n; i++ { m[i] = "foo" } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } ``` b797dc9088ad87976c76ab72c764bcc42a917496
1.0
Compare lists after operations, to require correctnes - Compare lists after operations, to require correctnes https://github.com/JonasMuehlmann/datastructures.go/blob/c5d510817d83387c67548f03bca14f73ab517343/maps/hashbidimap/hashbidimap_test.go#L340 ```golang package hashbidimap import ( "testing" "github.com/JonasMuehlmann/datastructures.go/tests" "github.com/JonasMuehlmann/datastructures.go/utils" "github.com/stretchr/testify/assert" ) func TestRemove(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] newMap *Map[string, int] toRemove string }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), newMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), toRemove: "foo", }, { name: "single item", toRemove: "foo", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item, target does not exist", toRemove: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", toRemove: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "baz": 3}), }, } for _, test := range tests { test.originalMap.Remove(utils.BasicComparator[string], test.toRemove) assert.ElementsMatchf(t, test.originalMap.GetKeys(), test.newMap.GetKeys(), test.name) } } func TestPut(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] newMap *Map[string, int] keyToAdd string valueToAdd int }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), keyToAdd: "foo", valueToAdd: 1, }, { name: "single item", keyToAdd: "foo", valueToAdd: 1, newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item, overwrite", keyToAdd: "foo", valueToAdd: 2, originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 2}), }, { name: "3 items", keyToAdd: "bar", valueToAdd: 2, originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "baz": 3}), newMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { test.originalMap.Put(test.keyToAdd, test.valueToAdd) assert.ElementsMatchf(t, test.originalMap.GetKeys(), test.newMap.GetKeys(), test.name) } } func TestGet(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] keyToGet string value int found bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), keyToGet: "foo", found: false, }, { name: "single item", keyToGet: "foo", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), value: 1, found: true, }, { name: "single item, target does not exist", keyToGet: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), found: false, }, { name: "3 items", keyToGet: "bar", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), value: 2, found: true, }, } for _, test := range tests { value, found := test.originalMap.Get(test.keyToGet) assert.Equalf(t, test.value, value, test.name) assert.Equalf(t, test.found, found, test.name) } } func TestGetKeys(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] keys []string }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), keys: []string{}, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), keys: []string{"foo"}, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), keys: []string{"foo", "bar", "baz"}, }, } for _, test := range tests { keys := test.originalMap.GetKeys() assert.ElementsMatchf(t, test.keys, keys, test.name) } } func TestGetValues(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] values []int }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), values: []int{}, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), values: []int{1}, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), values: []int{1, 2, 3}, }, } for _, test := range tests { values := test.originalMap.GetValues() assert.ElementsMatchf(t, test.values, values, test.name) } } func TestIsEmpty(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] isEmpty bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), isEmpty: true, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), isEmpty: false, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), isEmpty: false, }, } for _, test := range tests { isEmpty := test.originalMap.IsEmpty() assert.Equal(t, test.isEmpty, isEmpty, test.name) } } func TestClear(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] isEmptyBefore bool isEmptyAfter bool }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), isEmptyBefore: true, isEmptyAfter: true, }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), isEmptyBefore: false, isEmptyAfter: true, }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), isEmptyBefore: false, isEmptyAfter: true, }, } for _, test := range tests { isEmptyBefore := test.originalMap.IsEmpty() assert.Equal(t, test.isEmptyBefore, isEmptyBefore, test.name) test.originalMap.Clear() isEmptAfter := test.originalMap.IsEmpty() assert.Equal(t, test.isEmptyAfter, isEmptAfter, test.name) } } func TestNewFromIterator(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { it := test.originalMap.OrderedBegin(utils.BasicComparator[string]) newMap := NewFromIterator[string, int](utils.BasicComparator[string], utils.BasicComparator[int], it) assert.ElementsMatchf(t, test.originalMap.GetKeys(), newMap.GetKeys(), test.name) } } func TestNewFromIterators(t *testing.T) { tests := []struct { name string originalMap *Map[string, int] }{ { name: "empty list", originalMap: New[string, int](utils.BasicComparator[string], utils.BasicComparator[int]), }, { name: "single item", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1}), }, { name: "3 items", originalMap: NewFromMap[string, int](utils.BasicComparator[string], utils.BasicComparator[int], map[string]int{"foo": 1, "bar": 2, "baz": 3}), }, } for _, test := range tests { first := test.originalMap.OrderedBegin(utils.BasicComparator[string]) end := test.originalMap.OrderedEnd(utils.BasicComparator[string]) newMap := NewFromIterators[string, int](utils.BasicComparator[string], utils.BasicComparator[int], first, end) assert.ElementsMatchf(t, test.originalMap.GetKeys(), newMap.GetKeys(), test.name) } } // TODO: Compare lists after operations, to require correctnes func BenchmarkHashMapRemove(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) for i := 0; i < n; i++ { m.Put(i, "foo") } b.StartTimer() for i := 0; i < n; i++ { m.Remove(utils.BasicComparator[int], i) } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) for i := 0; i < n; i++ { m[i] = "foo" } b.StartTimer() for i := 0; i < n; i++ { delete(m, i) } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } func BenchmarkHashMapGet(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) for i := 0; i < n; i++ { m.Put(i, "foo") } b.StartTimer() for i := 0; i < n; i++ { _, _ = m.Get(i) } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) for i := 0; i < n; i++ { m[i] = "foo" } b.StartTimer() for i := 0; i < n; i++ { _, _ = m[i] } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } func BenchmarkHashMapPut(b *testing.B) { b.StopTimer() variants := []struct { name string f func(n int, name string) }{ { name: "Ours", f: func(n int, name string) { m := New[int, string](utils.BasicComparator[int], utils.BasicComparator[string]) b.StartTimer() for i := 0; i < n; i++ { m.Put(i, "foo") } b.StopTimer() }, }, { name: "Raw", f: func(n int, name string) { m := make(map[int]string) b.StartTimer() for i := 0; i < n; i++ { m[i] = "foo" } b.StopTimer() }, }, } for _, variant := range variants { tests.RunBenchmarkWithDefualtInputSizes(b, variant.name, variant.f) } } ``` b797dc9088ad87976c76ab72c764bcc42a917496
non_priority
compare lists after operations to require correctnes compare lists after operations to require correctnes golang package hashbidimap import testing github com jonasmuehlmann datastructures go tests github com jonasmuehlmann datastructures go utils github com stretchr testify assert func testremove t testing t tests struct name string originalmap map newmap map toremove string name empty list originalmap new utils basiccomparator utils basiccomparator newmap new utils basiccomparator utils basiccomparator toremove foo name single item toremove foo originalmap newfrommap utils basiccomparator utils basiccomparator map int foo newmap new utils basiccomparator utils basiccomparator name single item target does not exist toremove bar originalmap newfrommap utils basiccomparator utils basiccomparator map int foo newmap newfrommap utils basiccomparator utils basiccomparator map int foo name items toremove bar originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz newmap newfrommap utils basiccomparator utils basiccomparator map int foo baz for test range tests test originalmap remove utils basiccomparator test toremove assert elementsmatchf t test originalmap getkeys test newmap getkeys test name func testput t testing t tests struct name string originalmap map newmap map keytoadd string valuetoadd int name empty list originalmap new utils basiccomparator utils basiccomparator newmap newfrommap utils basiccomparator utils basiccomparator map int foo keytoadd foo valuetoadd name single item keytoadd foo valuetoadd newmap newfrommap utils basiccomparator utils basiccomparator map int foo originalmap new utils basiccomparator utils basiccomparator name single item overwrite keytoadd foo valuetoadd originalmap newfrommap utils basiccomparator utils basiccomparator map int foo newmap newfrommap utils basiccomparator utils basiccomparator map int foo name items keytoadd bar valuetoadd originalmap newfrommap utils basiccomparator utils basiccomparator map int foo baz newmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz for test range tests test originalmap put test keytoadd test valuetoadd assert elementsmatchf t test originalmap getkeys test newmap getkeys test name func testget t testing t tests struct name string originalmap map keytoget string value int found bool name empty list originalmap new utils basiccomparator utils basiccomparator keytoget foo found false name single item keytoget foo originalmap newfrommap utils basiccomparator utils basiccomparator map int foo value found true name single item target does not exist keytoget bar originalmap newfrommap utils basiccomparator utils basiccomparator map int foo found false name items keytoget bar originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz value found true for test range tests value found test originalmap get test keytoget assert equalf t test value value test name assert equalf t test found found test name func testgetkeys t testing t tests struct name string originalmap map keys string name empty list originalmap new utils basiccomparator utils basiccomparator keys string name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo keys string foo name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz keys string foo bar baz for test range tests keys test originalmap getkeys assert elementsmatchf t test keys keys test name func testgetvalues t testing t tests struct name string originalmap map values int name empty list originalmap new utils basiccomparator utils basiccomparator values int name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo values int name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz values int for test range tests values test originalmap getvalues assert elementsmatchf t test values values test name func testisempty t testing t tests struct name string originalmap map isempty bool name empty list originalmap new utils basiccomparator utils basiccomparator isempty true name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo isempty false name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz isempty false for test range tests isempty test originalmap isempty assert equal t test isempty isempty test name func testclear t testing t tests struct name string originalmap map isemptybefore bool isemptyafter bool name empty list originalmap new utils basiccomparator utils basiccomparator isemptybefore true isemptyafter true name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo isemptybefore false isemptyafter true name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz isemptybefore false isemptyafter true for test range tests isemptybefore test originalmap isempty assert equal t test isemptybefore isemptybefore test name test originalmap clear isemptafter test originalmap isempty assert equal t test isemptyafter isemptafter test name func testnewfromiterator t testing t tests struct name string originalmap map name empty list originalmap new utils basiccomparator utils basiccomparator name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz for test range tests it test originalmap orderedbegin utils basiccomparator newmap newfromiterator utils basiccomparator utils basiccomparator it assert elementsmatchf t test originalmap getkeys newmap getkeys test name func testnewfromiterators t testing t tests struct name string originalmap map name empty list originalmap new utils basiccomparator utils basiccomparator name single item originalmap newfrommap utils basiccomparator utils basiccomparator map int foo name items originalmap newfrommap utils basiccomparator utils basiccomparator map int foo bar baz for test range tests first test originalmap orderedbegin utils basiccomparator end test originalmap orderedend utils basiccomparator newmap newfromiterators utils basiccomparator utils basiccomparator first end assert elementsmatchf t test originalmap getkeys newmap getkeys test name todo compare lists after operations to require correctnes func benchmarkhashmapremove b testing b b stoptimer variants struct name string f func n int name string name ours f func n int name string m new utils basiccomparator utils basiccomparator for i i n i m put i foo b starttimer for i i n i m remove utils basiccomparator i b stoptimer name raw f func n int name string m make map string for i i n i m foo b starttimer for i i n i delete m i b stoptimer for variant range variants tests runbenchmarkwithdefualtinputsizes b variant name variant f func benchmarkhashmapget b testing b b stoptimer variants struct name string f func n int name string name ours f func n int name string m new utils basiccomparator utils basiccomparator for i i n i m put i foo b starttimer for i i n i m get i b stoptimer name raw f func n int name string m make map string for i i n i m foo b starttimer for i i n i m b stoptimer for variant range variants tests runbenchmarkwithdefualtinputsizes b variant name variant f func benchmarkhashmapput b testing b b stoptimer variants struct name string f func n int name string name ours f func n int name string m new utils basiccomparator utils basiccomparator b starttimer for i i n i m put i foo b stoptimer name raw f func n int name string m make map string b starttimer for i i n i m foo b stoptimer for variant range variants tests runbenchmarkwithdefualtinputsizes b variant name variant f
0
18,631
3,699,721,870
IssuesEvent
2016-02-29 02:32:30
allenhwkim/angularjs-google-maps
https://api.github.com/repos/allenhwkim/angularjs-google-maps
closed
test for controllers/map-controller.js
TEST
spec/controllers/map-controller_spec.js is missing. We need to unit test the following functions * addObject * deleteObjet * observeAttrSetObj
1.0
test for controllers/map-controller.js - spec/controllers/map-controller_spec.js is missing. We need to unit test the following functions * addObject * deleteObjet * observeAttrSetObj
non_priority
test for controllers map controller js spec controllers map controller spec js is missing we need to unit test the following functions addobject deleteobjet observeattrsetobj
0
281,529
30,888,882,591
IssuesEvent
2023-08-04 01:57:47
nidhi7598/linux-4.1.15_CVE-2019-10220
https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220
reopened
CVE-2022-1678 (High) detected in linuxlinux-4.4.302
Mend: dependency security vulnerability
## CVE-2022-1678 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/tcp_output.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux Kernel from 4.18 to 4.19, an improper update of sock reference in TCP pacing can lead to memory/netns leak, which can be used by remote clients. <p>Publish Date: 2022-05-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1678>CVE-2022-1678</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1678">https://www.linuxkernelcves.com/cves/CVE-2022-1678</a></p> <p>Release Date: 2022-05-25</p> <p>Fix Resolution: v4.20-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1678 (High) detected in linuxlinux-4.4.302 - ## CVE-2022-1678 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/tcp_output.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux Kernel from 4.18 to 4.19, an improper update of sock reference in TCP pacing can lead to memory/netns leak, which can be used by remote clients. <p>Publish Date: 2022-05-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1678>CVE-2022-1678</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1678">https://www.linuxkernelcves.com/cves/CVE-2022-1678</a></p> <p>Release Date: 2022-05-25</p> <p>Fix Resolution: v4.20-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net tcp output c vulnerability details an issue was discovered in the linux kernel from to an improper update of sock reference in tcp pacing can lead to memory netns leak which can be used by remote clients publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
29,651
5,659,281,973
IssuesEvent
2017-04-10 12:34:37
t9md/atom-vim-mode-plus
https://api.github.com/repos/t9md/atom-vim-mode-plus
closed
put-*-with-auto-indent breaks occurrence modifier for paragraphs
documentation Q&A
After mapping `p`/`P` to put-*-with-auto-indent, the occurrence modifier doesn't work in paragraphs, e.g. `cop` or `dop` has no effect. Without those mappings it works like expected. This is my keymap: ``` 'atom-text-editor.vim-mode-plus:not(.insert-mode)': 'P': 'vim-mode-plus:put-before-with-auto-indent' 'p': 'vim-mode-plus:put-after-with-auto-indent' ``` --- * Atom version: 1.15.0 * vim-mode-plus version: 0.88.0 * OS: Linux
1.0
put-*-with-auto-indent breaks occurrence modifier for paragraphs - After mapping `p`/`P` to put-*-with-auto-indent, the occurrence modifier doesn't work in paragraphs, e.g. `cop` or `dop` has no effect. Without those mappings it works like expected. This is my keymap: ``` 'atom-text-editor.vim-mode-plus:not(.insert-mode)': 'P': 'vim-mode-plus:put-before-with-auto-indent' 'p': 'vim-mode-plus:put-after-with-auto-indent' ``` --- * Atom version: 1.15.0 * vim-mode-plus version: 0.88.0 * OS: Linux
non_priority
put with auto indent breaks occurrence modifier for paragraphs after mapping p p to put with auto indent the occurrence modifier doesn t work in paragraphs e g cop or dop has no effect without those mappings it works like expected this is my keymap atom text editor vim mode plus not insert mode p vim mode plus put before with auto indent p vim mode plus put after with auto indent atom version vim mode plus version os linux
0
2,045
2,871,287,553
IssuesEvent
2015-06-08 00:24:14
SecUpwN/Android-IMSI-Catcher-Detector
https://api.github.com/repos/SecUpwN/Android-IMSI-Catcher-Detector
opened
Make Buildozer versions concide with app info display
BUG BUILD ENHANCEMENT
It's an annoying fact the the automatic development builds made by our [Buildozer](https://buildozer.io/) engine, doesn't coincide with the version that is shown in the app under `About AIMSICD`. I would very much like to see these version number coincide. That should also be true for the non-debug GitHub releases. This issue is closely related to #147, but that info is outdated in light of this one. Once resolved, we can close both. I've labelled this as a "bug" to prioritize it over other general enhancements.
1.0
Make Buildozer versions concide with app info display - It's an annoying fact the the automatic development builds made by our [Buildozer](https://buildozer.io/) engine, doesn't coincide with the version that is shown in the app under `About AIMSICD`. I would very much like to see these version number coincide. That should also be true for the non-debug GitHub releases. This issue is closely related to #147, but that info is outdated in light of this one. Once resolved, we can close both. I've labelled this as a "bug" to prioritize it over other general enhancements.
non_priority
make buildozer versions concide with app info display it s an annoying fact the the automatic development builds made by our engine doesn t coincide with the version that is shown in the app under about aimsicd i would very much like to see these version number coincide that should also be true for the non debug github releases this issue is closely related to but that info is outdated in light of this one once resolved we can close both i ve labelled this as a bug to prioritize it over other general enhancements
0
170,715
27,007,605,642
IssuesEvent
2023-02-10 12:59:46
jacanchaplais/graphicle
https://api.github.com/repos/jacanchaplais/graphicle
opened
Output FastJet clusters as a deque rather than a MaskGroup
enhancement design breaking-change
`MaskGroup` doesn't have any special useful functionality, it just makes things more complex with the string wrapped numerical indices. Return it as a deque, so that accessing the highest pT jets from the left is efficient.
1.0
Output FastJet clusters as a deque rather than a MaskGroup - `MaskGroup` doesn't have any special useful functionality, it just makes things more complex with the string wrapped numerical indices. Return it as a deque, so that accessing the highest pT jets from the left is efficient.
non_priority
output fastjet clusters as a deque rather than a maskgroup maskgroup doesn t have any special useful functionality it just makes things more complex with the string wrapped numerical indices return it as a deque so that accessing the highest pt jets from the left is efficient
0
41,667
5,383,974,905
IssuesEvent
2017-02-24 09:03:15
LiskHQ/lisk
https://api.github.com/repos/LiskHQ/lisk
closed
Travis build occasionally fails because of yarn installation error
test
Sometimes yarn installation process fails with error message: `error Received malformed response from registry. The registry may be down. info Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command. The command "yarn" failed and exited with 1 during .`
1.0
Travis build occasionally fails because of yarn installation error - Sometimes yarn installation process fails with error message: `error Received malformed response from registry. The registry may be down. info Visit https://yarnpkg.com/en/docs/cli/install for documentation about this command. The command "yarn" failed and exited with 1 during .`
non_priority
travis build occasionally fails because of yarn installation error sometimes yarn installation process fails with error message error received malformed response from registry the registry may be down info visit for documentation about this command the command yarn failed and exited with during
0
67,683
13,012,789,562
IssuesEvent
2020-07-25 07:51:21
jscastro76/threebox
https://api.github.com/repos/jscastro76/threebox
closed
Remove version logs from ColladaLoader and FBXLoader
:bookmark_tabs: code improvement
ColladaLoader and FBXLoader have a `consol.log` with the version of the file that is being processed. This is completely unnecessary.
1.0
Remove version logs from ColladaLoader and FBXLoader - ColladaLoader and FBXLoader have a `consol.log` with the version of the file that is being processed. This is completely unnecessary.
non_priority
remove version logs from colladaloader and fbxloader colladaloader and fbxloader have a consol log with the version of the file that is being processed this is completely unnecessary
0
14,746
4,973,933,617
IssuesEvent
2016-12-06 03:30:47
dart-lang/site-webdev
https://api.github.com/repos/dart-lang/site-webdev
closed
angular2 codelab TOC problem -- all entries are just "Use"
bug CodeLab
In the angular2 codelab autogenerated TOC all entries are just “Use” … that is because the `<foo>` element gets embedded as html. See the following image: ![screen shot 2016-11-14 at 1 11 08 pm](https://cloud.githubusercontent.com/assets/4140793/20283254/10298bc2-aa6d-11e6-9b9b-bb4a3518ea4d.png)
1.0
angular2 codelab TOC problem -- all entries are just "Use" - In the angular2 codelab autogenerated TOC all entries are just “Use” … that is because the `<foo>` element gets embedded as html. See the following image: ![screen shot 2016-11-14 at 1 11 08 pm](https://cloud.githubusercontent.com/assets/4140793/20283254/10298bc2-aa6d-11e6-9b9b-bb4a3518ea4d.png)
non_priority
codelab toc problem all entries are just use in the codelab autogenerated toc all entries are just “use” … that is because the element gets embedded as html see the following image
0
17,155
9,633,804,373
IssuesEvent
2019-05-15 19:35:59
MicrosoftDocs/sql-docs
https://api.github.com/repos/MicrosoftDocs/sql-docs
closed
sp_query_store_consistency_check doesn't exist in SQL 2016
doc-provided performance/tech sql/prod
"Verify Query Store is collecting query data continuously" section suggests running the `sp_query_store_consistency_check` procedure if the query store gets corrupted, but this sproc does not exist in SQL 2016, only SQL 2017 and later. See also: https://dba.stackexchange.com/questions/237521/trying-to-repair-query-store-but-sp-query-store-consistency-check-does-not-exis --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 155bfbc5-9c63-32c6-9d06-a9b3cdb4c60e * Version Independent ID: c854b2de-e509-e320-8151-67d690d486ed * Content: [Best Practice with the Query Store - SQL Server](https://docs.microsoft.com/en-us/sql/relational-databases/performance/best-practice-with-the-query-store?view=sql-server-2016#Verify) * Content Source: [docs/relational-databases/performance/best-practice-with-the-query-store.md](https://github.com/MicrosoftDocs/sql-docs/blob/live/docs/relational-databases/performance/best-practice-with-the-query-store.md) * Product: **sql** * Technology: **performance** * GitHub Login: @julieMSFT * Microsoft Alias: **jrasnick**
True
sp_query_store_consistency_check doesn't exist in SQL 2016 - "Verify Query Store is collecting query data continuously" section suggests running the `sp_query_store_consistency_check` procedure if the query store gets corrupted, but this sproc does not exist in SQL 2016, only SQL 2017 and later. See also: https://dba.stackexchange.com/questions/237521/trying-to-repair-query-store-but-sp-query-store-consistency-check-does-not-exis --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 155bfbc5-9c63-32c6-9d06-a9b3cdb4c60e * Version Independent ID: c854b2de-e509-e320-8151-67d690d486ed * Content: [Best Practice with the Query Store - SQL Server](https://docs.microsoft.com/en-us/sql/relational-databases/performance/best-practice-with-the-query-store?view=sql-server-2016#Verify) * Content Source: [docs/relational-databases/performance/best-practice-with-the-query-store.md](https://github.com/MicrosoftDocs/sql-docs/blob/live/docs/relational-databases/performance/best-practice-with-the-query-store.md) * Product: **sql** * Technology: **performance** * GitHub Login: @julieMSFT * Microsoft Alias: **jrasnick**
non_priority
sp query store consistency check doesn t exist in sql verify query store is collecting query data continuously section suggests running the sp query store consistency check procedure if the query store gets corrupted but this sproc does not exist in sql only sql and later see also document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product sql technology performance github login juliemsft microsoft alias jrasnick
0
50,322
6,354,250,302
IssuesEvent
2017-07-29 07:34:25
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
C# Design Notes for Jan 17, 2017
Area-Language Design Design Notes Language-C# New Language Feature - Pattern Matching New Language Feature - Tuples
# C# Language Design Notes for Jan 17, 2017 ## Agenda A few C# 7.0 issues to review. 1. Constant pattern semantics: which equality exactly? 2. Extension methods on tuples: should tuple conversions apply? # Constant pattern semantics Issue #16513 proposes a change to the semantics of constant patterns in `is` expressions. For the code ``` c# e is 42 ``` We currently generate the call `object.Equals(e, 42)` (or equivalent code), but we should instead generate `object.Equals(42, e)`. The implementation of `object.Equals` does a few reference equality and null checks, but otherwise delegates to the instance method `Equals` of its *first* argument. So with the current semantics the above would call `e.Equals(42)`, whereas in the proposal we would call `42.Equals(e)`. The issue lists several good reasons, and we can add more to the list: - The constant pattern isn't very *constant*, when it's behavior is determined by the non-constant operand! - Optimization opportunities are few when we cannot depend on known behavior of calling `c.Equals` on a constant value. - Intuitively, the pattern should do the testing, not the object being tested - Calling a method on the expression could cause side effects! - The difference from switch semantics is jarring - Switching would preserve the nice property of `is` expressions today that it only returns `true` if the left operand is implicitly convertible to the (type of the) right. There really is no downside to this, other than the little bit of work it requires to implement it. ## Conclusion Do it. # Extension methods on tuples Issue #16159 laments the facts that extension methods only apply to tuples if the tuple types match exactly. This is because extension methods currently only apply if there is an *identity, reference or boxing conversion* from the receiver to the type of the extension method's first parameter. The spirit of this rule is that if it applies to a type or its bases or interfaces, it will work. We agree that it *feels* like it should also work for tuples - at least "sometimes". We cannot make it just always work for tuple conversions, though, since they may recursively apply all kinds of conversions, including user defined conversions. We could check *recursively* through the tuple type for "the right kind of conversion". Compiler-wise this is a localized and low-risk change. It makes tuples compose well with extension methods. It's another place where things should "distribute over the elements" of the tuple. This is a now-or-never kind of change. It would be a breaking change to add later. ## Conclusion Try to do it now if at all possible.
2.0
C# Design Notes for Jan 17, 2017 - # C# Language Design Notes for Jan 17, 2017 ## Agenda A few C# 7.0 issues to review. 1. Constant pattern semantics: which equality exactly? 2. Extension methods on tuples: should tuple conversions apply? # Constant pattern semantics Issue #16513 proposes a change to the semantics of constant patterns in `is` expressions. For the code ``` c# e is 42 ``` We currently generate the call `object.Equals(e, 42)` (or equivalent code), but we should instead generate `object.Equals(42, e)`. The implementation of `object.Equals` does a few reference equality and null checks, but otherwise delegates to the instance method `Equals` of its *first* argument. So with the current semantics the above would call `e.Equals(42)`, whereas in the proposal we would call `42.Equals(e)`. The issue lists several good reasons, and we can add more to the list: - The constant pattern isn't very *constant*, when it's behavior is determined by the non-constant operand! - Optimization opportunities are few when we cannot depend on known behavior of calling `c.Equals` on a constant value. - Intuitively, the pattern should do the testing, not the object being tested - Calling a method on the expression could cause side effects! - The difference from switch semantics is jarring - Switching would preserve the nice property of `is` expressions today that it only returns `true` if the left operand is implicitly convertible to the (type of the) right. There really is no downside to this, other than the little bit of work it requires to implement it. ## Conclusion Do it. # Extension methods on tuples Issue #16159 laments the facts that extension methods only apply to tuples if the tuple types match exactly. This is because extension methods currently only apply if there is an *identity, reference or boxing conversion* from the receiver to the type of the extension method's first parameter. The spirit of this rule is that if it applies to a type or its bases or interfaces, it will work. We agree that it *feels* like it should also work for tuples - at least "sometimes". We cannot make it just always work for tuple conversions, though, since they may recursively apply all kinds of conversions, including user defined conversions. We could check *recursively* through the tuple type for "the right kind of conversion". Compiler-wise this is a localized and low-risk change. It makes tuples compose well with extension methods. It's another place where things should "distribute over the elements" of the tuple. This is a now-or-never kind of change. It would be a breaking change to add later. ## Conclusion Try to do it now if at all possible.
non_priority
c design notes for jan c language design notes for jan agenda a few c issues to review constant pattern semantics which equality exactly extension methods on tuples should tuple conversions apply constant pattern semantics issue proposes a change to the semantics of constant patterns in is expressions for the code c e is we currently generate the call object equals e or equivalent code but we should instead generate object equals e the implementation of object equals does a few reference equality and null checks but otherwise delegates to the instance method equals of its first argument so with the current semantics the above would call e equals whereas in the proposal we would call equals e the issue lists several good reasons and we can add more to the list the constant pattern isn t very constant when it s behavior is determined by the non constant operand optimization opportunities are few when we cannot depend on known behavior of calling c equals on a constant value intuitively the pattern should do the testing not the object being tested calling a method on the expression could cause side effects the difference from switch semantics is jarring switching would preserve the nice property of is expressions today that it only returns true if the left operand is implicitly convertible to the type of the right there really is no downside to this other than the little bit of work it requires to implement it conclusion do it extension methods on tuples issue laments the facts that extension methods only apply to tuples if the tuple types match exactly this is because extension methods currently only apply if there is an identity reference or boxing conversion from the receiver to the type of the extension method s first parameter the spirit of this rule is that if it applies to a type or its bases or interfaces it will work we agree that it feels like it should also work for tuples at least sometimes we cannot make it just always work for tuple conversions though since they may recursively apply all kinds of conversions including user defined conversions we could check recursively through the tuple type for the right kind of conversion compiler wise this is a localized and low risk change it makes tuples compose well with extension methods it s another place where things should distribute over the elements of the tuple this is a now or never kind of change it would be a breaking change to add later conclusion try to do it now if at all possible
0
107,660
13,494,623,739
IssuesEvent
2020-09-11 21:52:03
phetsims/circuit-construction-kit-common
https://api.github.com/repos/phetsims/circuit-construction-kit-common
opened
CCK: AC needs homescreen + navbar icons
design:artwork
CCK: AC needs homescreen + navbar icons It currently looks like this: ![image](https://user-images.githubusercontent.com/679486/92975973-b46aaf00-f446-11ea-8bff-d6fb4dc44df4.png)
1.0
CCK: AC needs homescreen + navbar icons - CCK: AC needs homescreen + navbar icons It currently looks like this: ![image](https://user-images.githubusercontent.com/679486/92975973-b46aaf00-f446-11ea-8bff-d6fb4dc44df4.png)
non_priority
cck ac needs homescreen navbar icons cck ac needs homescreen navbar icons it currently looks like this
0
3,965
4,128,315,616
IssuesEvent
2016-06-10 05:23:11
orientechnologies/orientdb
https://api.github.com/repos/orientechnologies/orientdb
closed
All Orient DB 2.1.15 threads are locked up for long periods of time
performance storage team waiting reply
java.util.concurrent.locks.ReentrantReadWriteLock$ReadLock.tryLock(long, TimeUnit) ReentrantReadWriteLock.java:871 com.orientechnologies.common.concur.resource.OSharedResourceAdaptive.acquireSharedLock() OSharedResourceAdaptive.java:142 com.orientechnologies.common.concur.resource.OSharedResourceAdaptiveExternal.acquireSharedLock() OSharedResourceAdaptiveExternal.java:44 com.orientechnologies.orient.core.storage.impl.local.OAbstractPaginatedStorage.readRecord(OCluster, ORecordId) OAbstractPaginatedStorage.java:1748 com.orientechnologies.orient.core.storage.impl.local.OAbstractPaginatedStorage.readRecord(ORecordId, String, boolean, ORecordCallback) OAbstractPaginatedStorage.java:780 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx$SimpleRecordReader.readRecord(OStorage, ORecordId, String, boolean, ORecordVersion) ODatabaseDocumentTx.java:3092 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.executeReadRecord(ORecordId, ORecord, ORecordVersion, String, boolean, boolean, boolean, OStorage$LOCKING_STRATEGY, ODatabaseDocumentTx$RecordReader) ODatabaseDocumentTx.java:1820 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean, boolean, boolean, OStorage$LOCKING_STRATEGY) OTransactionOptimistic.java:243 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean, boolean, OStorage$LOCKING_STRATEGY) OTransactionOptimistic.java:217 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean) OTransactionOptimistic.java:347 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.load(ORID) ODatabaseDocumentTx.java:1564 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.load(ORID) ODatabaseDocumentTx.java:122 com.orientechnologies.orient.core.id.ORecordId.getRecord() ORecordId.java:324 com.tinkerpop.blueprints.impls.orient.OrientElementIterator.next() OrientElementIterator.java:62 com.tinkerpop.blueprints.impls.orient.OrientElementIterator.next() OrientElementIterator.java:35 com.stimulus.archiva.database.blueprints.BluePrintDatabase.f(Database$ItemType, Database$IndexedProperty, String) MailArchiva:383 com.stimulus.archiva.database.blueprints.BluePrintDatabase.d(Database$ItemType, String) MailArchiva:367 com.stimulus.archiva.database.blueprints.BluePrintDatabase.c(Database$ItemType, String) MailArchiva:594
True
All Orient DB 2.1.15 threads are locked up for long periods of time - java.util.concurrent.locks.ReentrantReadWriteLock$ReadLock.tryLock(long, TimeUnit) ReentrantReadWriteLock.java:871 com.orientechnologies.common.concur.resource.OSharedResourceAdaptive.acquireSharedLock() OSharedResourceAdaptive.java:142 com.orientechnologies.common.concur.resource.OSharedResourceAdaptiveExternal.acquireSharedLock() OSharedResourceAdaptiveExternal.java:44 com.orientechnologies.orient.core.storage.impl.local.OAbstractPaginatedStorage.readRecord(OCluster, ORecordId) OAbstractPaginatedStorage.java:1748 com.orientechnologies.orient.core.storage.impl.local.OAbstractPaginatedStorage.readRecord(ORecordId, String, boolean, ORecordCallback) OAbstractPaginatedStorage.java:780 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx$SimpleRecordReader.readRecord(OStorage, ORecordId, String, boolean, ORecordVersion) ODatabaseDocumentTx.java:3092 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.executeReadRecord(ORecordId, ORecord, ORecordVersion, String, boolean, boolean, boolean, OStorage$LOCKING_STRATEGY, ODatabaseDocumentTx$RecordReader) ODatabaseDocumentTx.java:1820 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean, boolean, boolean, OStorage$LOCKING_STRATEGY) OTransactionOptimistic.java:243 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean, boolean, OStorage$LOCKING_STRATEGY) OTransactionOptimistic.java:217 com.orientechnologies.orient.core.tx.OTransactionOptimistic.loadRecord(ORID, ORecord, String, boolean) OTransactionOptimistic.java:347 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.load(ORID) ODatabaseDocumentTx.java:1564 com.orientechnologies.orient.core.db.document.ODatabaseDocumentTx.load(ORID) ODatabaseDocumentTx.java:122 com.orientechnologies.orient.core.id.ORecordId.getRecord() ORecordId.java:324 com.tinkerpop.blueprints.impls.orient.OrientElementIterator.next() OrientElementIterator.java:62 com.tinkerpop.blueprints.impls.orient.OrientElementIterator.next() OrientElementIterator.java:35 com.stimulus.archiva.database.blueprints.BluePrintDatabase.f(Database$ItemType, Database$IndexedProperty, String) MailArchiva:383 com.stimulus.archiva.database.blueprints.BluePrintDatabase.d(Database$ItemType, String) MailArchiva:367 com.stimulus.archiva.database.blueprints.BluePrintDatabase.c(Database$ItemType, String) MailArchiva:594
non_priority
all orient db threads are locked up for long periods of time java util concurrent locks reentrantreadwritelock readlock trylock long timeunit reentrantreadwritelock java com orientechnologies common concur resource osharedresourceadaptive acquiresharedlock osharedresourceadaptive java com orientechnologies common concur resource osharedresourceadaptiveexternal acquiresharedlock osharedresourceadaptiveexternal java com orientechnologies orient core storage impl local oabstractpaginatedstorage readrecord ocluster orecordid oabstractpaginatedstorage java com orientechnologies orient core storage impl local oabstractpaginatedstorage readrecord orecordid string boolean orecordcallback oabstractpaginatedstorage java com orientechnologies orient core db document odatabasedocumenttx simplerecordreader readrecord ostorage orecordid string boolean orecordversion odatabasedocumenttx java com orientechnologies orient core db document odatabasedocumenttx executereadrecord orecordid orecord orecordversion string boolean boolean boolean ostorage locking strategy odatabasedocumenttx recordreader odatabasedocumenttx java com orientechnologies orient core tx otransactionoptimistic loadrecord orid orecord string boolean boolean boolean ostorage locking strategy otransactionoptimistic java com orientechnologies orient core tx otransactionoptimistic loadrecord orid orecord string boolean boolean ostorage locking strategy otransactionoptimistic java com orientechnologies orient core tx otransactionoptimistic loadrecord orid orecord string boolean otransactionoptimistic java com orientechnologies orient core db document odatabasedocumenttx load orid odatabasedocumenttx java com orientechnologies orient core db document odatabasedocumenttx load orid odatabasedocumenttx java com orientechnologies orient core id orecordid getrecord orecordid java com tinkerpop blueprints impls orient orientelementiterator next orientelementiterator java com tinkerpop blueprints impls orient orientelementiterator next orientelementiterator java com stimulus archiva database blueprints blueprintdatabase f database itemtype database indexedproperty string mailarchiva com stimulus archiva database blueprints blueprintdatabase d database itemtype string mailarchiva com stimulus archiva database blueprints blueprintdatabase c database itemtype string mailarchiva
0
21,137
3,462,206,298
IssuesEvent
2015-12-20 18:57:08
prettydiff/prettydiff
https://api.github.com/repos/prettydiff/prettydiff
closed
Ternary problems in JavaScript beautification
Defect Parsing QA
This defect is mangling beautification in JavaScript, but it is not corrupting parsing or code output... so it isn't anything critical. I found some defects when examining the code from #233 . I have solved most of the problems created a new unit test file for the cases discovered so far: /test/samples_raw/beautification_javascript_aatern.txt The only remaining problem (that I am currently aware of) is this guy function () { return val ? typeofe ? elem ? name : a : valspecified ? valvalue : b : c; } Which becomes function () { return val ? typeofe ? elem ? name : a : valspecified ? valvalue : b : c; }
1.0
Ternary problems in JavaScript beautification - This defect is mangling beautification in JavaScript, but it is not corrupting parsing or code output... so it isn't anything critical. I found some defects when examining the code from #233 . I have solved most of the problems created a new unit test file for the cases discovered so far: /test/samples_raw/beautification_javascript_aatern.txt The only remaining problem (that I am currently aware of) is this guy function () { return val ? typeofe ? elem ? name : a : valspecified ? valvalue : b : c; } Which becomes function () { return val ? typeofe ? elem ? name : a : valspecified ? valvalue : b : c; }
non_priority
ternary problems in javascript beautification this defect is mangling beautification in javascript but it is not corrupting parsing or code output so it isn t anything critical i found some defects when examining the code from i have solved most of the problems created a new unit test file for the cases discovered so far test samples raw beautification javascript aatern txt the only remaining problem that i am currently aware of is this guy function return val typeofe elem name a valspecified valvalue b c which becomes function return val typeofe elem name a valspecified valvalue b c
0
95,953
12,065,709,850
IssuesEvent
2020-04-16 10:25:52
Lenkly/festivent
https://api.github.com/repos/Lenkly/festivent
opened
Add pages for the app
backend design frontend
The app needs several pages for Navigation and Autonavigation. As a User I'd like to "test" the app feature without a need to register. I can add my name, my taste in music and receive a list of suggestions with detailed view. As a User I would also like to create an account to save festivals to my favourite list. I want to change my information and switch between app themes in the settings.
1.0
Add pages for the app - The app needs several pages for Navigation and Autonavigation. As a User I'd like to "test" the app feature without a need to register. I can add my name, my taste in music and receive a list of suggestions with detailed view. As a User I would also like to create an account to save festivals to my favourite list. I want to change my information and switch between app themes in the settings.
non_priority
add pages for the app the app needs several pages for navigation and autonavigation as a user i d like to test the app feature without a need to register i can add my name my taste in music and receive a list of suggestions with detailed view as a user i would also like to create an account to save festivals to my favourite list i want to change my information and switch between app themes in the settings
0
210,218
16,091,376,930
IssuesEvent
2021-04-26 17:07:36
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
UnexpectedBehaviour during PVC creation error
TestCase failing scale
Full log: https://ocs4-jenkins-csb-ocsqe.apps.ocp4.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/559/testReport/tests.e2e.scale.test_pvc_creation_deletion_scale/TestPVCCreationDeletionScale/test_all_4_type_pvc_creation_deletion_scale/ logging.info("Waited for more than 3mins still no data") > raise UnexpectedBehaviour( f"There is no pvc creation data in CSI logs for {no_data_list}" )
1.0
UnexpectedBehaviour during PVC creation error - Full log: https://ocs4-jenkins-csb-ocsqe.apps.ocp4.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/559/testReport/tests.e2e.scale.test_pvc_creation_deletion_scale/TestPVCCreationDeletionScale/test_all_4_type_pvc_creation_deletion_scale/ logging.info("Waited for more than 3mins still no data") > raise UnexpectedBehaviour( f"There is no pvc creation data in CSI logs for {no_data_list}" )
non_priority
unexpectedbehaviour during pvc creation error full log logging info waited for more than still no data raise unexpectedbehaviour f there is no pvc creation data in csi logs for no data list
0
7,163
10,310,453,132
IssuesEvent
2019-08-29 15:11:52
heim-rs/heim
https://api.github.com/repos/heim-rs/heim
closed
process::Process::parent method
A-process C-enhancement O-linux O-macos O-windows
Should be as simple as combination of `Process::parent_pid` and `Process::get`
1.0
process::Process::parent method - Should be as simple as combination of `Process::parent_pid` and `Process::get`
non_priority
process process parent method should be as simple as combination of process parent pid and process get
0
78,699
15,051,812,177
IssuesEvent
2021-02-03 14:31:56
ices-eg/DIG
https://api.github.com/repos/ices-eg/DIG
closed
Two new metiers for InterCatch
Approved-AsIs vocab: CodeValue vocab: new
Data submitter for WKWEST (Claire Moore) requested two metiers to be added (see below) to InterCatch for bycatch fisheries for stock pol.27,67 Metier 1: LHP_DEF_0_0_0_all  -  description: Hand-lines and hand-operated pole and lines, Demersal fish Metier 2: LHP_SPF_0_0_0_all  -  description: Hand-lines and hand-operated pole and lines, Small pelagic fish
1.0
Two new metiers for InterCatch - Data submitter for WKWEST (Claire Moore) requested two metiers to be added (see below) to InterCatch for bycatch fisheries for stock pol.27,67 Metier 1: LHP_DEF_0_0_0_all  -  description: Hand-lines and hand-operated pole and lines, Demersal fish Metier 2: LHP_SPF_0_0_0_all  -  description: Hand-lines and hand-operated pole and lines, Small pelagic fish
non_priority
two new metiers for intercatch data submitter for wkwest claire moore requested two metiers to be added see below to intercatch for bycatch fisheries for stock pol metier lhp def all    description hand lines and hand operated pole and lines demersal fish metier lhp spf all    description hand lines and hand operated pole and lines small pelagic fish
0
286,536
31,652,451,432
IssuesEvent
2023-09-07 00:19:01
samq-wsdemo/vuln_rust
https://api.github.com/repos/samq-wsdemo/vuln_rust
opened
CVE-2022-37434 (Critical) detected in libz-sys-1.1.2.crate
Mend: dependency security vulnerability
## CVE-2022-37434 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libz-sys-1.1.2.crate</b></p></summary> <p>Low-level bindings to the system libz library (also known as zlib). </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/libz-sys/1.1.2/download">https://crates.io/api/v1/crates/libz-sys/1.1.2/download</a></p> <p>Path to dependency file: /Cargo.toml</p> <p>Path to vulnerable library: /home/wss-scanner/.cargo/registry/cache/github.com-1ecc6299db9ec823/libz-sys-1.1.2.crate</p> <p> Dependency Hierarchy: - chttp-0.5.5.crate (Root Library) - curl-0.4.35.crate - curl-sys-0.4.41+curl-7.75.0.crate - :x: **libz-sys-1.1.2.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> zlib through 1.2.12 has a heap-based buffer over-read or buffer overflow in inflate in inflate.c via a large gzip header extra field. NOTE: only applications that call inflateGetHeader are affected. Some common applications bundle the affected zlib source code but may be unable to call inflateGetHeader (e.g., see the nodejs/node reference). <p>Publish Date: 2022-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37434>CVE-2022-37434</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-37434 (Critical) detected in libz-sys-1.1.2.crate - ## CVE-2022-37434 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libz-sys-1.1.2.crate</b></p></summary> <p>Low-level bindings to the system libz library (also known as zlib). </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/libz-sys/1.1.2/download">https://crates.io/api/v1/crates/libz-sys/1.1.2/download</a></p> <p>Path to dependency file: /Cargo.toml</p> <p>Path to vulnerable library: /home/wss-scanner/.cargo/registry/cache/github.com-1ecc6299db9ec823/libz-sys-1.1.2.crate</p> <p> Dependency Hierarchy: - chttp-0.5.5.crate (Root Library) - curl-0.4.35.crate - curl-sys-0.4.41+curl-7.75.0.crate - :x: **libz-sys-1.1.2.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> zlib through 1.2.12 has a heap-based buffer over-read or buffer overflow in inflate in inflate.c via a large gzip header extra field. NOTE: only applications that call inflateGetHeader are affected. Some common applications bundle the affected zlib source code but may be unable to call inflateGetHeader (e.g., see the nodejs/node reference). <p>Publish Date: 2022-08-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37434>CVE-2022-37434</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_priority
cve critical detected in libz sys crate cve critical severity vulnerability vulnerable library libz sys crate low level bindings to the system libz library also known as zlib library home page a href path to dependency file cargo toml path to vulnerable library home wss scanner cargo registry cache github com libz sys crate dependency hierarchy chttp crate root library curl crate curl sys curl crate x libz sys crate vulnerable library found in base branch master vulnerability details zlib through has a heap based buffer over read or buffer overflow in inflate in inflate c via a large gzip header extra field note only applications that call inflategetheader are affected some common applications bundle the affected zlib source code but may be unable to call inflategetheader e g see the nodejs node reference publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href
0
74,151
7,382,478,758
IssuesEvent
2018-03-15 05:10:02
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Rancher Windows in Feb Release
Epic area/windows kind/feature status/to-test team/cn
This is the main issue for Rancher Windows support in Feb Release. There following major features for this release. * Rancher Agent upgrade * Webhook service for windows * CLI support for Windows environment * Windows environment validation test
1.0
Rancher Windows in Feb Release - This is the main issue for Rancher Windows support in Feb Release. There following major features for this release. * Rancher Agent upgrade * Webhook service for windows * CLI support for Windows environment * Windows environment validation test
non_priority
rancher windows in feb release this is the main issue for rancher windows support in feb release there following major features for this release rancher agent upgrade webhook service for windows cli support for windows environment windows environment validation test
0
158,846
12,426,773,047
IssuesEvent
2020-05-24 23:05:26
JMRI/JMRI
https://api.github.com/repos/JMRI/JMRI
opened
Failure to load NCE-Specific Startup actions
Add to Next Test Release Bug
Current HEAD doesn't load my NCE status actions: ![Current ](https://user-images.githubusercontent.com/13841457/82766868-ff47c300-9e65-11ea-84eb-12c08d9b3b1f.png) V4.9.16: ![V4-19-6](https://user-images.githubusercontent.com/13841457/82766853-dfb09a80-9e65-11ea-8740-6fd2ab7bd7b4.png) The build log flags a depreciation warning: ``` /Users/heap/jmri/java/src/apps/startup/StartupActionModelUtil.java:188: warning: [deprecation] StartupActionFactory in apps.startup has been deprecated asLoader.forEach(factory -> addActions(factory)); ``` [NetBeans Log.txt](https://github.com/JMRI/JMRI/files/4674579/NetBeans.Log.txt) Another example: ![Example 2 V4-19-6](https://user-images.githubusercontent.com/13841457/82766940-9f9de780-9e66-11ea-9d54-7779ba32721c.png) ![Example 2 Current](https://user-images.githubusercontent.com/13841457/82766947-a75d8c00-9e66-11ea-8d6c-fff3e986d335.png) [NCE_Power_Pro_Bare_copy.jmri.zip](https://github.com/JMRI/JMRI/files/4674581/NCE_Power_Pro_Bare_copy.jmri.zip) [NCE_Power_Pro.jmri.zip](https://github.com/JMRI/JMRI/files/4674582/NCE_Power_Pro.jmri.zip)
1.0
Failure to load NCE-Specific Startup actions - Current HEAD doesn't load my NCE status actions: ![Current ](https://user-images.githubusercontent.com/13841457/82766868-ff47c300-9e65-11ea-84eb-12c08d9b3b1f.png) V4.9.16: ![V4-19-6](https://user-images.githubusercontent.com/13841457/82766853-dfb09a80-9e65-11ea-8740-6fd2ab7bd7b4.png) The build log flags a depreciation warning: ``` /Users/heap/jmri/java/src/apps/startup/StartupActionModelUtil.java:188: warning: [deprecation] StartupActionFactory in apps.startup has been deprecated asLoader.forEach(factory -> addActions(factory)); ``` [NetBeans Log.txt](https://github.com/JMRI/JMRI/files/4674579/NetBeans.Log.txt) Another example: ![Example 2 V4-19-6](https://user-images.githubusercontent.com/13841457/82766940-9f9de780-9e66-11ea-9d54-7779ba32721c.png) ![Example 2 Current](https://user-images.githubusercontent.com/13841457/82766947-a75d8c00-9e66-11ea-8d6c-fff3e986d335.png) [NCE_Power_Pro_Bare_copy.jmri.zip](https://github.com/JMRI/JMRI/files/4674581/NCE_Power_Pro_Bare_copy.jmri.zip) [NCE_Power_Pro.jmri.zip](https://github.com/JMRI/JMRI/files/4674582/NCE_Power_Pro.jmri.zip)
non_priority
failure to load nce specific startup actions current head doesn t load my nce status actions the build log flags a depreciation warning users heap jmri java src apps startup startupactionmodelutil java warning startupactionfactory in apps startup has been deprecated asloader foreach factory addactions factory another example
0
64,072
26,605,767,155
IssuesEvent
2023-01-23 19:12:29
microsoft/BotFramework-DirectLineJS
https://api.github.com/repos/microsoft/BotFramework-DirectLineJS
closed
DirectlineStreaming doesn't work with Node.js
bug customer-reported Bot Services customer-replied-to
When trying to replace Directline with DirectlineStreaming, the websocket connection fails. The failure is in [botframework-streaming/lib/webSocket/nodeWebSocketClient](https://github.com/microsoft/botbuilder-js/blob/main/libraries/botframework-streaming/src/webSocket/nodeWebSocketClient.ts#L74). `this._url` is the full URL (in my case that's wss://my-streaming-bot.azurewebsites.net/.bot/v3/directline/conversations/connect?token=eyJh...), and not the host alone, so websocket initialization throws an exception. Moreover, the error message is not reflected to the calling code. It just writes "Unable to connect client to Node transport.". Minimal example follows. ```javascript import crypto from 'crypto'; import XMLHttpRequest from 'xhr2'; import WebSocket from 'ws'; global.XMLHttpRequest = XMLHttpRequest; global.WebSocket = WebSocket; import { DirectLineStreaming, ConnectionStatus } from 'botframework-directlinejs'; import axios from 'axios'; const botUrl = 'https://my-streaming-bot.azurewebsites.net/.bot/v3/directline'; const userId = crypto.randomUUID(); const username = `dl_ac_test-${userId.substring(0, userId.indexOf('-'))}`; const botSecret = 'XXXYYYZZZ'; async function main() { const data = { user: { id: username } }; const token = (await axios.post('tokens/generate', data, { baseURL: botUrl, headers: { Authorization: `Bearer ${botSecret}` } })).data; const directLine = new DirectLineStreaming({ conversationId: token.conversationId, token: token.token, domain: botUrl, webSocket: true, }); directLine.connectionStatus$.subscribe( (connectionStatus) => console.log('DirectLine status: ' + ConnectionStatus[connectionStatus]) ); directLine.activity$.subscribe({ next(activity) { console.log(activity); } }); } main(); ```
1.0
DirectlineStreaming doesn't work with Node.js - When trying to replace Directline with DirectlineStreaming, the websocket connection fails. The failure is in [botframework-streaming/lib/webSocket/nodeWebSocketClient](https://github.com/microsoft/botbuilder-js/blob/main/libraries/botframework-streaming/src/webSocket/nodeWebSocketClient.ts#L74). `this._url` is the full URL (in my case that's wss://my-streaming-bot.azurewebsites.net/.bot/v3/directline/conversations/connect?token=eyJh...), and not the host alone, so websocket initialization throws an exception. Moreover, the error message is not reflected to the calling code. It just writes "Unable to connect client to Node transport.". Minimal example follows. ```javascript import crypto from 'crypto'; import XMLHttpRequest from 'xhr2'; import WebSocket from 'ws'; global.XMLHttpRequest = XMLHttpRequest; global.WebSocket = WebSocket; import { DirectLineStreaming, ConnectionStatus } from 'botframework-directlinejs'; import axios from 'axios'; const botUrl = 'https://my-streaming-bot.azurewebsites.net/.bot/v3/directline'; const userId = crypto.randomUUID(); const username = `dl_ac_test-${userId.substring(0, userId.indexOf('-'))}`; const botSecret = 'XXXYYYZZZ'; async function main() { const data = { user: { id: username } }; const token = (await axios.post('tokens/generate', data, { baseURL: botUrl, headers: { Authorization: `Bearer ${botSecret}` } })).data; const directLine = new DirectLineStreaming({ conversationId: token.conversationId, token: token.token, domain: botUrl, webSocket: true, }); directLine.connectionStatus$.subscribe( (connectionStatus) => console.log('DirectLine status: ' + ConnectionStatus[connectionStatus]) ); directLine.activity$.subscribe({ next(activity) { console.log(activity); } }); } main(); ```
non_priority
directlinestreaming doesn t work with node js when trying to replace directline with directlinestreaming the websocket connection fails the failure is in this url is the full url in my case that s wss my streaming bot azurewebsites net bot directline conversations connect token eyjh and not the host alone so websocket initialization throws an exception moreover the error message is not reflected to the calling code it just writes unable to connect client to node transport minimal example follows javascript import crypto from crypto import xmlhttprequest from import websocket from ws global xmlhttprequest xmlhttprequest global websocket websocket import directlinestreaming connectionstatus from botframework directlinejs import axios from axios const boturl const userid crypto randomuuid const username dl ac test userid substring userid indexof const botsecret xxxyyyzzz async function main const data user id username const token await axios post tokens generate data baseurl boturl headers authorization bearer botsecret data const directline new directlinestreaming conversationid token conversationid token token token domain boturl websocket true directline connectionstatus subscribe connectionstatus console log directline status connectionstatus directline activity subscribe next activity console log activity main
0
93,414
26,942,116,077
IssuesEvent
2023-02-08 03:42:54
sagemath/sage
https://api.github.com/repos/sagemath/sage
closed
Unable to build from source due to setuptools error with old-style SPKGs
t: bug c: build c: python3
<div id="comment:0"></div> I'm trying to compile Sage 9.6 from the tarball on Ubuntu 20.04. I have installed all the recommend packages from APT (although `./configure` complains about some of them not being installed; I've tried the recommended `sudo apt-get install` command but it didn't help). Running `./configure --prefix=$HOME/.local --with-python=$HOME/.local/bin/python` works fine, but then when running `make`, I get an error. It says `setuptools-59.8.0` failed to build, and gives me the following log: ``` cat: /build/pkgs/setuptools/type: No such file or directory Error: Installing old-style SPKGs is no longer supported ``` I have no idea what the issue might be, but my relevant environment variables are as follows: `SAGE_ROOT=$HOME/.local/src/sage-9.6 MAKE="make -j4" SAGE_CHECK="yes" SAGE_DEBUG="no"`. I'm looking for advice rather than a patch. Component: **python3** Keywords: **setuptools, install, make** _Issue created by migration from https://trac.sagemath.org/ticket/34332_
1.0
Unable to build from source due to setuptools error with old-style SPKGs - <div id="comment:0"></div> I'm trying to compile Sage 9.6 from the tarball on Ubuntu 20.04. I have installed all the recommend packages from APT (although `./configure` complains about some of them not being installed; I've tried the recommended `sudo apt-get install` command but it didn't help). Running `./configure --prefix=$HOME/.local --with-python=$HOME/.local/bin/python` works fine, but then when running `make`, I get an error. It says `setuptools-59.8.0` failed to build, and gives me the following log: ``` cat: /build/pkgs/setuptools/type: No such file or directory Error: Installing old-style SPKGs is no longer supported ``` I have no idea what the issue might be, but my relevant environment variables are as follows: `SAGE_ROOT=$HOME/.local/src/sage-9.6 MAKE="make -j4" SAGE_CHECK="yes" SAGE_DEBUG="no"`. I'm looking for advice rather than a patch. Component: **python3** Keywords: **setuptools, install, make** _Issue created by migration from https://trac.sagemath.org/ticket/34332_
non_priority
unable to build from source due to setuptools error with old style spkgs i m trying to compile sage from the tarball on ubuntu i have installed all the recommend packages from apt although configure complains about some of them not being installed i ve tried the recommended sudo apt get install command but it didn t help running configure prefix home local with python home local bin python works fine but then when running make i get an error it says setuptools failed to build and gives me the following log cat build pkgs setuptools type no such file or directory error installing old style spkgs is no longer supported i have no idea what the issue might be but my relevant environment variables are as follows sage root home local src sage make make sage check yes sage debug no i m looking for advice rather than a patch component keywords setuptools install make issue created by migration from
0
61,826
6,759,133,754
IssuesEvent
2017-10-24 16:07:08
RIOT-OS/RIOT
https://api.github.com/repos/RIOT-OS/RIOT
closed
tests/posix_semaphore: fails
native tests
While testing on 2017.10-RC1 with native the following error showed up. > ######################### TEST4: > first: sem_init s1 > first: wait 1 sec for s1 > first: timed out > first: waited only 1000105 usec => FAILED > ######################### DONE > Timeout in expect script
1.0
tests/posix_semaphore: fails - While testing on 2017.10-RC1 with native the following error showed up. > ######################### TEST4: > first: sem_init s1 > first: wait 1 sec for s1 > first: timed out > first: waited only 1000105 usec => FAILED > ######################### DONE > Timeout in expect script
non_priority
tests posix semaphore fails while testing on with native the following error showed up first sem init first wait sec for first timed out first waited only usec failed done timeout in expect script
0
273,158
20,774,722,774
IssuesEvent
2022-03-16 09:25:27
cucumber/cucumber-js
https://api.github.com/repos/cucumber/cucumber-js
closed
Incomplete CHANGELOG
:book: documentation
Could you please list the breaking changes in the formatter which are mentioned in CHANGELOG? > Custom formatters: Formatter helpers and EventDataCollector had breaking changes to support retry Currently, it is really hard to figure out what changed and what must be changed to support Cucumber 6 if we create a custom formatter. E.g. `getTestCaseData` has been removed and it is not clear, how [getTestCaseAttempt](https://github.com/cucumber/cucumber-js/blob/master/src/formatter/helpers/event_data_collector.ts#L47) would replace it, or what other function does. At the same time the link to `EventDataCollector` in the [documentation](https://github.com/cucumber/cucumber-js/blob/master/docs/custom_formatters.md) is not valid.
1.0
Incomplete CHANGELOG - Could you please list the breaking changes in the formatter which are mentioned in CHANGELOG? > Custom formatters: Formatter helpers and EventDataCollector had breaking changes to support retry Currently, it is really hard to figure out what changed and what must be changed to support Cucumber 6 if we create a custom formatter. E.g. `getTestCaseData` has been removed and it is not clear, how [getTestCaseAttempt](https://github.com/cucumber/cucumber-js/blob/master/src/formatter/helpers/event_data_collector.ts#L47) would replace it, or what other function does. At the same time the link to `EventDataCollector` in the [documentation](https://github.com/cucumber/cucumber-js/blob/master/docs/custom_formatters.md) is not valid.
non_priority
incomplete changelog could you please list the breaking changes in the formatter which are mentioned in changelog custom formatters formatter helpers and eventdatacollector had breaking changes to support retry currently it is really hard to figure out what changed and what must be changed to support cucumber if we create a custom formatter e g gettestcasedata has been removed and it is not clear how would replace it or what other function does at the same time the link to eventdatacollector in the is not valid
0
179,174
14,693,239,232
IssuesEvent
2021-01-03 07:45:57
apache/incubator-doris
https://api.github.com/repos/apache/incubator-doris
closed
Some errors in the document need to be rectified
kind/documentation
Hi, all. I found some mistakes in our document, such as spelling errors、default values error. For example: 1. spelling mistakes: `docs\zh-CN\sql-reference\sql-statements\Administration\ADMIN SET CONFIG.md`: 可设置的配置项,可以通过 `AMDIN` SHOW FRONTEND CONFIG; 命令查看。 `AMDIN` should be `ADMIN` 2. default value mistakes: `docs\zh-CN\administrator-guide\operation\tablet-repair-and-balance.md`: disable_balance default value should be false And so on. I want to correct these mistakes. I will submit my changes soon.
1.0
Some errors in the document need to be rectified - Hi, all. I found some mistakes in our document, such as spelling errors、default values error. For example: 1. spelling mistakes: `docs\zh-CN\sql-reference\sql-statements\Administration\ADMIN SET CONFIG.md`: 可设置的配置项,可以通过 `AMDIN` SHOW FRONTEND CONFIG; 命令查看。 `AMDIN` should be `ADMIN` 2. default value mistakes: `docs\zh-CN\administrator-guide\operation\tablet-repair-and-balance.md`: disable_balance default value should be false And so on. I want to correct these mistakes. I will submit my changes soon.
non_priority
some errors in the document need to be rectified hi all i found some mistakes in our document such as spelling errors、default values error for example spelling mistakes docs zh cn sql reference sql statements administration admin set config md 可设置的配置项,可以通过 amdin show frontend config 命令查看。 amdin should be admin default value mistakes docs zh cn administrator guide operation tablet repair and balance md disable balance default value should be false and so on i want to correct these mistakes i will submit my changes soon
0
231,267
17,671,660,429
IssuesEvent
2021-08-23 07:08:42
oasis-open/csaf-documentation
https://api.github.com/repos/oasis-open/csaf-documentation
opened
Add a news section
enhancement website documentation
Please add a sections with news about CSAF (e.g. that it is currently in Public Review).
1.0
Add a news section - Please add a sections with news about CSAF (e.g. that it is currently in Public Review).
non_priority
add a news section please add a sections with news about csaf e g that it is currently in public review
0
172,550
21,050,500,343
IssuesEvent
2022-03-31 20:06:52
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
opened
Vulnerability roundup 112: nomad-1.1.8: 4 advisories [7.5]
1.severity: security
[search](https://search.nix.gsc.io/?q=nomad&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=nomad+in%3Apath&type=Code) * [ ] [CVE-2022-24683](https://nvd.nist.gov/vuln/detail/CVE-2022-24683) CVSSv3=7.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24685](https://nvd.nist.gov/vuln/detail/CVE-2022-24685) CVSSv3=7.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24684](https://nvd.nist.gov/vuln/detail/CVE-2022-24684) CVSSv3=6.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24686](https://nvd.nist.gov/vuln/detail/CVE-2022-24686) CVSSv3=5.9 (nixos-21.11, nixos-unstable) ## CVE details ### CVE-2022-24683 HashiCorp Nomad and Nomad Enterprise 0.9.2 through 1.0.17, 1.1.11, and 1.2.5 allow operators with read-fs and alloc-exec (or job-submit) capabilities to read arbitrary files on the host filesystem as root. ### CVE-2022-24685 HashiCorp Nomad and Nomad Enterprise 1.x before 1.0.17, 1.1.x before 1.1.12, and 1.2.x before 1.2.6 has Uncontrolled Resource Consumption. ### CVE-2022-24684 HashiCorp Nomad and Nomad Enterprise before 1.0.17, 1.1.x before 1.1.12, and 1.2.x before 1.2.6 has Uncontrolled Resource Consumption. ### CVE-2022-24686 HashiCorp Nomad and Nomad Enterprise 0.3.0 through 1.0.17, 1.1.11, and 1.2.5 artifact download functionality has a race condition such that the Nomad client agent could download the wrong artifact into the wrong destination. Fixed in 1.0.18, 1.1.12, and 1.2.6 ----- Scanned versions: nixos-21.11: efea022d6fe; nixos-unstable: 710fed5a248. Cc @endocrimes Cc @kalbasit Cc @maxeaubrey Cc @pradeepchhetri Cc @rushmorem Cc @techknowlogick
True
Vulnerability roundup 112: nomad-1.1.8: 4 advisories [7.5] - [search](https://search.nix.gsc.io/?q=nomad&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=nomad+in%3Apath&type=Code) * [ ] [CVE-2022-24683](https://nvd.nist.gov/vuln/detail/CVE-2022-24683) CVSSv3=7.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24685](https://nvd.nist.gov/vuln/detail/CVE-2022-24685) CVSSv3=7.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24684](https://nvd.nist.gov/vuln/detail/CVE-2022-24684) CVSSv3=6.5 (nixos-21.11, nixos-unstable) * [ ] [CVE-2022-24686](https://nvd.nist.gov/vuln/detail/CVE-2022-24686) CVSSv3=5.9 (nixos-21.11, nixos-unstable) ## CVE details ### CVE-2022-24683 HashiCorp Nomad and Nomad Enterprise 0.9.2 through 1.0.17, 1.1.11, and 1.2.5 allow operators with read-fs and alloc-exec (or job-submit) capabilities to read arbitrary files on the host filesystem as root. ### CVE-2022-24685 HashiCorp Nomad and Nomad Enterprise 1.x before 1.0.17, 1.1.x before 1.1.12, and 1.2.x before 1.2.6 has Uncontrolled Resource Consumption. ### CVE-2022-24684 HashiCorp Nomad and Nomad Enterprise before 1.0.17, 1.1.x before 1.1.12, and 1.2.x before 1.2.6 has Uncontrolled Resource Consumption. ### CVE-2022-24686 HashiCorp Nomad and Nomad Enterprise 0.3.0 through 1.0.17, 1.1.11, and 1.2.5 artifact download functionality has a race condition such that the Nomad client agent could download the wrong artifact into the wrong destination. Fixed in 1.0.18, 1.1.12, and 1.2.6 ----- Scanned versions: nixos-21.11: efea022d6fe; nixos-unstable: 710fed5a248. Cc @endocrimes Cc @kalbasit Cc @maxeaubrey Cc @pradeepchhetri Cc @rushmorem Cc @techknowlogick
non_priority
vulnerability roundup nomad advisories nixos nixos unstable nixos nixos unstable nixos nixos unstable nixos nixos unstable cve details cve hashicorp nomad and nomad enterprise through and allow operators with read fs and alloc exec or job submit capabilities to read arbitrary files on the host filesystem as root cve hashicorp nomad and nomad enterprise x before x before and x before has uncontrolled resource consumption cve hashicorp nomad and nomad enterprise before x before and x before has uncontrolled resource consumption cve hashicorp nomad and nomad enterprise through and artifact download functionality has a race condition such that the nomad client agent could download the wrong artifact into the wrong destination fixed in and scanned versions nixos nixos unstable cc endocrimes cc kalbasit cc maxeaubrey cc pradeepchhetri cc rushmorem cc techknowlogick
0
152,109
23,915,590,385
IssuesEvent
2022-09-09 12:24:52
tarantool/tt
https://api.github.com/repos/tarantool/tt
closed
tt import/tt export
teamE in design
```bash tt import name -format csv -space users users.csv tt import name -format csv -space users < file.csv tt export name -format csv -space users users.csv tt export name -format csv -space users > users.csv ``` Options: - `--format` - file format (CSV is default) - `--space` - space name for import/export. Required option - `--match name1=csvrow1:name2=csvrow2:...` - Match for CSV and space fields **for import and export**. - csvrow is CSV row number (from 1 to N) - name - is field name or field number - `--header=on|off` - (`off` - default). - `on` - export fieldnames to the first CSV line, import fieldnames from the first CSV line (can raise error) - `off` - the first CSV line is a data line You could make alias `--match header` and `--header on`. Import (only) options: - `--progress {file}` - use the file to store import progress If the option "`match`" option doesn't contain any primary key fields, the utility will generate a primary key for each record: autoincrement for `numeric` fields, UUID-string for `string` fields, `UUID` for `UUID` fields. - `--failed {skip|error}` - behaviour with error tuples (example duplicate key or constraint errors) - `--update {new|all}` - update exists records or create new only. (The option means something if PK is present in input CSV)
1.0
tt import/tt export - ```bash tt import name -format csv -space users users.csv tt import name -format csv -space users < file.csv tt export name -format csv -space users users.csv tt export name -format csv -space users > users.csv ``` Options: - `--format` - file format (CSV is default) - `--space` - space name for import/export. Required option - `--match name1=csvrow1:name2=csvrow2:...` - Match for CSV and space fields **for import and export**. - csvrow is CSV row number (from 1 to N) - name - is field name or field number - `--header=on|off` - (`off` - default). - `on` - export fieldnames to the first CSV line, import fieldnames from the first CSV line (can raise error) - `off` - the first CSV line is a data line You could make alias `--match header` and `--header on`. Import (only) options: - `--progress {file}` - use the file to store import progress If the option "`match`" option doesn't contain any primary key fields, the utility will generate a primary key for each record: autoincrement for `numeric` fields, UUID-string for `string` fields, `UUID` for `UUID` fields. - `--failed {skip|error}` - behaviour with error tuples (example duplicate key or constraint errors) - `--update {new|all}` - update exists records or create new only. (The option means something if PK is present in input CSV)
non_priority
tt import tt export bash tt import name format csv space users users csv tt import name format csv space users file csv tt export name format csv space users users csv tt export name format csv space users users csv options format file format csv is default space space name for import export required option match match for csv and space fields for import and export csvrow is csv row number from to n name is field name or field number header on off off default on export fieldnames to the first csv line import fieldnames from the first csv line can raise error off the first csv line is a data line you could make alias match header and header on import only options progress file use the file to store import progress if the option match option doesn t contain any primary key fields the utility will generate a primary key for each record autoincrement for numeric fields uuid string for string fields uuid for uuid fields failed skip error behaviour with error tuples example duplicate key or constraint errors update new all update exists records or create new only the option means something if pk is present in input csv
0
159,351
20,048,363,708
IssuesEvent
2022-02-03 01:09:23
kapseliboi/mimic
https://api.github.com/repos/kapseliboi/mimic
opened
CVE-2020-7788 (High) detected in ini-1.3.4.tgz, ini-1.3.5.tgz
security vulnerability
## CVE-2020-7788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ini-1.3.4.tgz</b>, <b>ini-1.3.5.tgz</b></p></summary> <p> <details><summary><b>ini-1.3.4.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.4.tgz">https://registry.npmjs.org/ini/-/ini-1.3.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ini/package.json</p> <p> Dependency Hierarchy: - js-beautify-1.6.14.tgz (Root Library) - config-chain-1.1.11.tgz - :x: **ini-1.3.4.tgz** (Vulnerable Library) </details> <details><summary><b>ini-1.3.5.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - rc-1.2.8.tgz - :x: **ini-1.3.5.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution (ini): 1.3.6</p> <p>Direct dependency fix Resolution (js-beautify): 1.7.0</p><p>Fix Resolution (ini): 1.3.6</p> <p>Direct dependency fix Resolution (karma): 1.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7788 (High) detected in ini-1.3.4.tgz, ini-1.3.5.tgz - ## CVE-2020-7788 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ini-1.3.4.tgz</b>, <b>ini-1.3.5.tgz</b></p></summary> <p> <details><summary><b>ini-1.3.4.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.4.tgz">https://registry.npmjs.org/ini/-/ini-1.3.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ini/package.json</p> <p> Dependency Hierarchy: - js-beautify-1.6.14.tgz (Root Library) - config-chain-1.1.11.tgz - :x: **ini-1.3.4.tgz** (Vulnerable Library) </details> <details><summary><b>ini-1.3.5.tgz</b></p></summary> <p>An ini encoder/decoder for node</p> <p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - rc-1.2.8.tgz - :x: **ini-1.3.5.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution (ini): 1.3.6</p> <p>Direct dependency fix Resolution (js-beautify): 1.7.0</p><p>Fix Resolution (ini): 1.3.6</p> <p>Direct dependency fix Resolution (karma): 1.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in ini tgz ini tgz cve high severity vulnerability vulnerable libraries ini tgz ini tgz ini tgz an ini encoder decoder for node library home page a href path to dependency file package json path to vulnerable library node modules ini package json dependency hierarchy js beautify tgz root library config chain tgz x ini tgz vulnerable library ini tgz an ini encoder decoder for node library home page a href dependency hierarchy karma tgz root library chokidar tgz fsevents tgz node pre gyp tgz rc tgz x ini tgz vulnerable library found in base branch master vulnerability details this affects the package ini before if an attacker submits a malicious ini file to an application that parses it with ini parse they will pollute the prototype on the application this can be exploited further depending on the context publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ini direct dependency fix resolution js beautify fix resolution ini direct dependency fix resolution karma step up your open source security game with whitesource
0
74,951
15,386,651,107
IssuesEvent
2021-03-03 08:30:59
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] Updated name of external Connector is not reflecting on update button at case
Team: SecuritySolution Team:Threat Hunting bug fixed impact:medium v7.12.0
**Describe the bug** Updated name of external Connector is not reflecting on update button at case **Build Details:** ``` Platform : Production Version : 7.12.0 BC2 Commit : 4f65a5a1268fa78f1af9117d12312e1cee433376 Build: 39000 ``` **Browser Details** All **Preconditions** 1.Elastic Cloud environment having version 7.12.0 should be available. 2.Audit beat should be installed on above environment. 3.IBM Connector should be created on above environment. let say connector name would be "**ksingh_ibm_resilient connector**" ![image](https://user-images.githubusercontent.com/59917825/109001145-93330d80-76ca-11eb-9bba-89686d77d332.png) **Steps to Reproduce** 1. Navigate to Kibana and Select Security Solution App from the left navigation panel. 2. Go to Case tab and add above IBM connector to it. 3. Go to Edit External Connector again and update the existing name of IBM connector. let sat "**ksingh_ibm_resilient connector updated**" ![image](https://user-images.githubusercontent.com/59917825/109000569-e2c50980-76c9-11eb-9113-dea52946977f.png) 4. Again go to created case. 5. Observed that name of connector was not updated to current updated name. ![image](https://user-images.githubusercontent.com/59917825/109001003-62eb6f00-76ca-11eb-8452-e4ad118447d8.png) **Actual Result** Updated name of external Connector is not reflecting on update button at case **Expected Result** Updated name of external Connector should reflect on update button at case **Whats Working** - N/A **Whats Not Working** - N/A **Screenshots** ![image](https://user-images.githubusercontent.com/59917825/109000582-e9538100-76c9-11eb-86f1-3b1d9012bedc.png) **logs** N/A
True
[Security Solution] Updated name of external Connector is not reflecting on update button at case - **Describe the bug** Updated name of external Connector is not reflecting on update button at case **Build Details:** ``` Platform : Production Version : 7.12.0 BC2 Commit : 4f65a5a1268fa78f1af9117d12312e1cee433376 Build: 39000 ``` **Browser Details** All **Preconditions** 1.Elastic Cloud environment having version 7.12.0 should be available. 2.Audit beat should be installed on above environment. 3.IBM Connector should be created on above environment. let say connector name would be "**ksingh_ibm_resilient connector**" ![image](https://user-images.githubusercontent.com/59917825/109001145-93330d80-76ca-11eb-9bba-89686d77d332.png) **Steps to Reproduce** 1. Navigate to Kibana and Select Security Solution App from the left navigation panel. 2. Go to Case tab and add above IBM connector to it. 3. Go to Edit External Connector again and update the existing name of IBM connector. let sat "**ksingh_ibm_resilient connector updated**" ![image](https://user-images.githubusercontent.com/59917825/109000569-e2c50980-76c9-11eb-9113-dea52946977f.png) 4. Again go to created case. 5. Observed that name of connector was not updated to current updated name. ![image](https://user-images.githubusercontent.com/59917825/109001003-62eb6f00-76ca-11eb-8452-e4ad118447d8.png) **Actual Result** Updated name of external Connector is not reflecting on update button at case **Expected Result** Updated name of external Connector should reflect on update button at case **Whats Working** - N/A **Whats Not Working** - N/A **Screenshots** ![image](https://user-images.githubusercontent.com/59917825/109000582-e9538100-76c9-11eb-86f1-3b1d9012bedc.png) **logs** N/A
non_priority
updated name of external connector is not reflecting on update button at case describe the bug updated name of external connector is not reflecting on update button at case build details platform production version commit build browser details all preconditions elastic cloud environment having version should be available audit beat should be installed on above environment ibm connector should be created on above environment let say connector name would be ksingh ibm resilient connector steps to reproduce navigate to kibana and select security solution app from the left navigation panel go to case tab and add above ibm connector to it go to edit external connector again and update the existing name of ibm connector let sat ksingh ibm resilient connector updated again go to created case observed that name of connector was not updated to current updated name actual result updated name of external connector is not reflecting on update button at case expected result updated name of external connector should reflect on update button at case whats working n a whats not working n a screenshots logs n a
0
50,317
7,583,121,076
IssuesEvent
2018-04-25 07:48:17
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
document makeWrapper
9.needs: documentation
There is no documentation whatsoever about one of the most basic utilities in nixpkgs. There is exactly one very short source code comment about usage.
1.0
document makeWrapper - There is no documentation whatsoever about one of the most basic utilities in nixpkgs. There is exactly one very short source code comment about usage.
non_priority
document makewrapper there is no documentation whatsoever about one of the most basic utilities in nixpkgs there is exactly one very short source code comment about usage
0
193,073
22,216,050,625
IssuesEvent
2022-06-08 01:50:47
artsking/linux-4.1.15
https://api.github.com/repos/artsking/linux-4.1.15
reopened
CVE-2019-15917 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2019-15917 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/bluetooth/hci_ldisc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/bluetooth/hci_ldisc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.5. There is a use-after-free issue when hci_uart_register_dev() fails in hci_uart_set_proto() in drivers/bluetooth/hci_ldisc.c. <p>Publish Date: 2019-09-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15917>CVE-2019-15917</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15917">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15917</a></p> <p>Release Date: 2020-03-09</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15917 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2019-15917 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/bluetooth/hci_ldisc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/bluetooth/hci_ldisc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.5. There is a use-after-free issue when hci_uart_register_dev() fails in hci_uart_set_proto() in drivers/bluetooth/hci_ldisc.c. <p>Publish Date: 2019-09-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15917>CVE-2019-15917</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15917">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15917</a></p> <p>Release Date: 2020-03-09</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers bluetooth hci ldisc c drivers bluetooth hci ldisc c vulnerability details an issue was discovered in the linux kernel before there is a use after free issue when hci uart register dev fails in hci uart set proto in drivers bluetooth hci ldisc c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
116,649
11,938,599,836
IssuesEvent
2020-04-02 14:01:17
kaiyuanshe/osschat
https://api.github.com/repos/kaiyuanshe/osschat
closed
[DOC] - Create a static documentation site based on Readthedocs
documentation enhancement
Description : - Create a static documentation site based on Readthedocs. - Support markdowm documents with the Recommonmark module. - Automatically triggering builds with Readthedocs' webhook. - Use Readthedocs to bind our custom domain, such as `doc.oss.chat`.
1.0
[DOC] - Create a static documentation site based on Readthedocs - Description : - Create a static documentation site based on Readthedocs. - Support markdowm documents with the Recommonmark module. - Automatically triggering builds with Readthedocs' webhook. - Use Readthedocs to bind our custom domain, such as `doc.oss.chat`.
non_priority
create a static documentation site based on readthedocs description create a static documentation site based on readthedocs support markdowm documents with the recommonmark module automatically triggering builds with readthedocs webhook use readthedocs to bind our custom domain such as doc oss chat
0
20,310
4,536,253,876
IssuesEvent
2016-09-08 19:51:12
odlgroup/odl
https://api.github.com/repos/odlgroup/odl
opened
Migrate examples to use functionals
documentation odl solvers
Now that functionals are soon in the old examples need to be migrated so that they use the new and nicer functional syntax.
1.0
Migrate examples to use functionals - Now that functionals are soon in the old examples need to be migrated so that they use the new and nicer functional syntax.
non_priority
migrate examples to use functionals now that functionals are soon in the old examples need to be migrated so that they use the new and nicer functional syntax
0
417,791
28,111,111,183
IssuesEvent
2023-03-31 07:15:28
axmszr/ped
https://api.github.com/repos/axmszr/ped
opened
Incorrect command name for Delete
type.DocumentationBug severity.Medium
Delete is implemented as `del`, but the User Guide still sometimes shows `delete` spelt out in full (which does not work). ![image.png](https://raw.githubusercontent.com/axmszr/ped/main/files/5d1d4c8b-c59d-42b2-aadd-9962b54d963d.png) <!--session: 1680242647550-b023bf1e-340f-4a31-8b4d-5e392a9f6496--> <!--Version: Web v3.4.7-->
1.0
Incorrect command name for Delete - Delete is implemented as `del`, but the User Guide still sometimes shows `delete` spelt out in full (which does not work). ![image.png](https://raw.githubusercontent.com/axmszr/ped/main/files/5d1d4c8b-c59d-42b2-aadd-9962b54d963d.png) <!--session: 1680242647550-b023bf1e-340f-4a31-8b4d-5e392a9f6496--> <!--Version: Web v3.4.7-->
non_priority
incorrect command name for delete delete is implemented as del but the user guide still sometimes shows delete spelt out in full which does not work
0
76,386
9,436,194,529
IssuesEvent
2019-04-13 03:59:58
team-quick/quick-rails
https://api.github.com/repos/team-quick/quick-rails
opened
기본 뼈대 디자인 의견 수렴
design enhancement
<img width="1203" alt="스크린샷 2019-04-13 오후 12 58 13" src="https://user-images.githubusercontent.com/21256346/56074364-e90aaf00-5deb-11e9-98f4-1d53ee60435e.png"> 기본적인 스케치는 일단 이렇게 해 봤는데 색상과 배치에 관한 의견을 수렴합니다
1.0
기본 뼈대 디자인 의견 수렴 - <img width="1203" alt="스크린샷 2019-04-13 오후 12 58 13" src="https://user-images.githubusercontent.com/21256346/56074364-e90aaf00-5deb-11e9-98f4-1d53ee60435e.png"> 기본적인 스케치는 일단 이렇게 해 봤는데 색상과 배치에 관한 의견을 수렴합니다
non_priority
기본 뼈대 디자인 의견 수렴 img width alt 스크린샷 오후 src 기본적인 스케치는 일단 이렇게 해 봤는데 색상과 배치에 관한 의견을 수렴합니다
0
180,160
14,740,251,008
IssuesEvent
2021-01-07 08:46:47
eclipse/capella
https://api.github.com/repos/eclipse/capella
closed
Refer to Eclipse diff/merge and patterns doc
documentation enhancement
An update of the doc of the Diff/Merge and Modeling Patterns components is required. - Diff/Merge: the current doc mostly duplicates the doc that is now embedded in EMF Diff/Merge; it should refer to it instead. - Patterns: the current doc is very small and lacks a reference to the Eclipse complete guide in the EDM Patterns doc. `🆔 ECLIPSE-554829 / POLARSYS-1793` `👷 olivier.constant` `📅 2017-10-17` `🔎 1.2.0`
1.0
Refer to Eclipse diff/merge and patterns doc - An update of the doc of the Diff/Merge and Modeling Patterns components is required. - Diff/Merge: the current doc mostly duplicates the doc that is now embedded in EMF Diff/Merge; it should refer to it instead. - Patterns: the current doc is very small and lacks a reference to the Eclipse complete guide in the EDM Patterns doc. `🆔 ECLIPSE-554829 / POLARSYS-1793` `👷 olivier.constant` `📅 2017-10-17` `🔎 1.2.0`
non_priority
refer to eclipse diff merge and patterns doc an update of the doc of the diff merge and modeling patterns components is required diff merge the current doc mostly duplicates the doc that is now embedded in emf diff merge it should refer to it instead patterns the current doc is very small and lacks a reference to the eclipse complete guide in the edm patterns doc 🆔 eclipse polarsys 👷 olivier constant 📅 🔎
0
53,447
7,839,509,705
IssuesEvent
2018-06-18 13:53:51
jackadull/jackadull-related
https://api.github.com/repos/jackadull/jackadull-related
opened
Add Coveralls Badges
documentation
After completing #10, integrated Coveralls badges in all projects. Also depends on jackadull/jackadocs#3.
1.0
Add Coveralls Badges - After completing #10, integrated Coveralls badges in all projects. Also depends on jackadull/jackadocs#3.
non_priority
add coveralls badges after completing integrated coveralls badges in all projects also depends on jackadull jackadocs
0
296,687
25,569,513,755
IssuesEvent
2022-11-30 16:35:23
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
opened
Small Text Bug - Wrong UpdateNotification in console
Requires Testing
### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug I tried the FastAsyncWorldEdit version attached to the last Github-Release and one maked as Release from Jenkins and both times FAWE prints following line in my server console: ``` [17:29:00 WARN]: [com.fastasyncworldedit.core.util.UpdateNotification] You are using a snapshot or a custom version of FAWE. This is not an official build distributed via https://www.spigotmc.org/resources/13932/ ``` ### To Reproduce 1. Use a FAWE version from a release 2. Look into console ### Expected behaviour No information about an SNAPSHOT or CUSTOM version, because it isn't one. ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/c75a209c226542c6bfbf7db4a4278722 ### Fawe Version FastAsyncWorldEdit version 2.4.10 ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists. ### Anything else? _No response_
1.0
Small Text Bug - Wrong UpdateNotification in console - ### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug I tried the FastAsyncWorldEdit version attached to the last Github-Release and one maked as Release from Jenkins and both times FAWE prints following line in my server console: ``` [17:29:00 WARN]: [com.fastasyncworldedit.core.util.UpdateNotification] You are using a snapshot or a custom version of FAWE. This is not an official build distributed via https://www.spigotmc.org/resources/13932/ ``` ### To Reproduce 1. Use a FAWE version from a release 2. Look into console ### Expected behaviour No information about an SNAPSHOT or CUSTOM version, because it isn't one. ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/c75a209c226542c6bfbf7db4a4278722 ### Fawe Version FastAsyncWorldEdit version 2.4.10 ### Checklist - [X] I have included a Fawe debugpaste. - [X] I am using the newest build from https://ci.athion.net/job/FastAsyncWorldEdit/ and the issue still persists. ### Anything else? _No response_
non_priority
small text bug wrong updatenotification in console server implementation paper server version describe the bug i tried the fastasyncworldedit version attached to the last github release and one maked as release from jenkins and both times fawe prints following line in my server console you are using a snapshot or a custom version of fawe this is not an official build distributed via to reproduce use a fawe version from a release look into console expected behaviour no information about an snapshot or custom version because it isn t one screenshots videos no response error log if applicable no response fawe debugpaste fawe version fastasyncworldedit version checklist i have included a fawe debugpaste i am using the newest build from and the issue still persists anything else no response
0
29,471
8,360,507,030
IssuesEvent
2018-10-03 11:47:22
inkss/inkss.github.io
https://api.github.com/repos/inkss/inkss.github.io
closed
Hexo 博客搭建
/2018/09/15/hexo-blog-build/ Gitalk
https://inkss.cn/2018/09/15/hexo-blog-build/ 本文偏向于记录备份,非新手向教程,仅供参考。 🐱先说搭建这个博客的原因:在搭建博客之前我的所有的文章都是存储在 inkss/markdown 仓库中,直到某天 在投简历的时候发现有的要求填写博客链接 ,掐指一算在玩腻饥荒后似乎还闲置着一个腾讯云的学生主机,索性自己折腾一个博客出来。而以简单快速为目的话,还是主推: Hexo 。Hexo 可以帮忙生成全静态的网页,所以在存储位置上可以有:Githu
1.0
Hexo 博客搭建 - https://inkss.cn/2018/09/15/hexo-blog-build/ 本文偏向于记录备份,非新手向教程,仅供参考。 🐱先说搭建这个博客的原因:在搭建博客之前我的所有的文章都是存储在 inkss/markdown 仓库中,直到某天 在投简历的时候发现有的要求填写博客链接 ,掐指一算在玩腻饥荒后似乎还闲置着一个腾讯云的学生主机,索性自己折腾一个博客出来。而以简单快速为目的话,还是主推: Hexo 。Hexo 可以帮忙生成全静态的网页,所以在存储位置上可以有:Githu
non_priority
hexo 博客搭建 本文偏向于记录备份,非新手向教程,仅供参考。 🐱先说搭建这个博客的原因:在搭建博客之前我的所有的文章都是存储在 inkss markdown 仓库中,直到某天 在投简历的时候发现有的要求填写博客链接 ,掐指一算在玩腻饥荒后似乎还闲置着一个腾讯云的学生主机,索性自己折腾一个博客出来。而以简单快速为目的话,还是主推: hexo 。hexo 可以帮忙生成全静态的网页,所以在存储位置上可以有:githu
0
51,315
13,635,093,066
IssuesEvent
2020-09-25 01:52:20
nasifimtiazohi/openmrs-core-2.3.0
https://api.github.com/repos/nasifimtiazohi/openmrs-core-2.3.0
opened
CVE-2019-14379 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2019-14379 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: openmrs-core-2.3.0/webapp/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles default typing when ehcache is used (because of net.sf.ehcache.transaction.manager.DefaultTransactionManagerLookup), leading to remote code execution. <p>Publish Date: 2019-07-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14379>CVE-2019-14379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14379">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14379</a></p> <p>Release Date: 2019-07-29</p> <p>Fix Resolution: 2.9.9.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14379 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2019-14379 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: openmrs-core-2.3.0/webapp/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles default typing when ehcache is used (because of net.sf.ehcache.transaction.manager.DefaultTransactionManagerLookup), leading to remote code execution. <p>Publish Date: 2019-07-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14379>CVE-2019-14379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14379">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14379</a></p> <p>Release Date: 2019-07-29</p> <p>Fix Resolution: 2.9.9.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file openmrs core webapp pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details subtypevalidator java in fasterxml jackson databind before mishandles default typing when ehcache is used because of net sf ehcache transaction manager defaulttransactionmanagerlookup leading to remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
287,920
24,874,210,149
IssuesEvent
2022-10-27 17:37:11
ecadlabs/taquito
https://api.github.com/repos/ecadlabs/taquito
closed
extend test coverage for beacon-wallet package using mocks
testing🧪
Additional test coverage for the beacon wallet package in Taquito will require setting up mocks. Get the coverage to green if possible,
1.0
extend test coverage for beacon-wallet package using mocks - Additional test coverage for the beacon wallet package in Taquito will require setting up mocks. Get the coverage to green if possible,
non_priority
extend test coverage for beacon wallet package using mocks additional test coverage for the beacon wallet package in taquito will require setting up mocks get the coverage to green if possible
0
44,315
9,560,258,150
IssuesEvent
2019-05-03 19:03:55
Microsoft/azure-pipelines-image-generation
https://api.github.com/repos/Microsoft/azure-pipelines-image-generation
closed
Services: Add WinAppDriver (Windows)
area:Testing/code coverage image:Windows issue:Enhancement stale
## Tool info - Tool name: WinAppDriver - Add or update? add - Desired version: 1.1 - If this is an add request: - Brief description of tool: drive automated UI tests of Windows applications - URL for tool's homepage: http://github.com/microsoft/winappdriver - Open source license type (Apache, MIT, GPL, etc.): MIT ## Areas and ecosystems <!--- This helps us route to the right owner: check one or two, don't worry about getting it perfect --> - [ ] Android - [ ] C / C++ - [ ] Container tools - [ ] Database servers or clients - [ ] General purpose tools - [ ] Go - [ ] Java - [ ] Javascript, Node.js, and TypeScript - [ ] iOS, macOS, and Apple-specific tools - [ ] .NET Core or .NET Desktop - [ ] PHP - [ ] Python - [ ] Ruby - [ ] Rust - [ ] Scala - [X ] Testing and code coverage - [ ] Xamarin - [ ] Other ## Hosted pools <!--- Where do you want to see this tool? Select all that apply --> - [ ] macOS - [ ] Ubuntu - [X ] Windows
1.0
Services: Add WinAppDriver (Windows) - ## Tool info - Tool name: WinAppDriver - Add or update? add - Desired version: 1.1 - If this is an add request: - Brief description of tool: drive automated UI tests of Windows applications - URL for tool's homepage: http://github.com/microsoft/winappdriver - Open source license type (Apache, MIT, GPL, etc.): MIT ## Areas and ecosystems <!--- This helps us route to the right owner: check one or two, don't worry about getting it perfect --> - [ ] Android - [ ] C / C++ - [ ] Container tools - [ ] Database servers or clients - [ ] General purpose tools - [ ] Go - [ ] Java - [ ] Javascript, Node.js, and TypeScript - [ ] iOS, macOS, and Apple-specific tools - [ ] .NET Core or .NET Desktop - [ ] PHP - [ ] Python - [ ] Ruby - [ ] Rust - [ ] Scala - [X ] Testing and code coverage - [ ] Xamarin - [ ] Other ## Hosted pools <!--- Where do you want to see this tool? Select all that apply --> - [ ] macOS - [ ] Ubuntu - [X ] Windows
non_priority
services add winappdriver windows tool info tool name winappdriver add or update add desired version if this is an add request brief description of tool drive automated ui tests of windows applications url for tool s homepage open source license type apache mit gpl etc mit areas and ecosystems android c c container tools database servers or clients general purpose tools go java javascript node js and typescript ios macos and apple specific tools net core or net desktop php python ruby rust scala testing and code coverage xamarin other hosted pools macos ubuntu windows
0
13,306
3,700,961,603
IssuesEvent
2016-02-29 10:58:13
playframework/playframework
https://api.github.com/repos/playframework/playframework
closed
Document usage of encryptAES / decryptAES
community documentation
There's nowhere in our docs where we give an example of encrypting/decrypting data. I think it could be helpful to share an example of when and how we might deal with encrypted data. Also, something else to mention in the migration guide potentially is that in 2.4 when the encryption algorithm changed, it seems to have changed what characters can appear in the encrypted string. I ran into an issue where we assumed encrypted data would be alphanumeric, which appears to be a true of AES. In Play 2.4 when we changed to AES/CTR/NoPadding we see new characters like -, =, and +.
1.0
Document usage of encryptAES / decryptAES - There's nowhere in our docs where we give an example of encrypting/decrypting data. I think it could be helpful to share an example of when and how we might deal with encrypted data. Also, something else to mention in the migration guide potentially is that in 2.4 when the encryption algorithm changed, it seems to have changed what characters can appear in the encrypted string. I ran into an issue where we assumed encrypted data would be alphanumeric, which appears to be a true of AES. In Play 2.4 when we changed to AES/CTR/NoPadding we see new characters like -, =, and +.
non_priority
document usage of encryptaes decryptaes there s nowhere in our docs where we give an example of encrypting decrypting data i think it could be helpful to share an example of when and how we might deal with encrypted data also something else to mention in the migration guide potentially is that in when the encryption algorithm changed it seems to have changed what characters can appear in the encrypted string i ran into an issue where we assumed encrypted data would be alphanumeric which appears to be a true of aes in play when we changed to aes ctr nopadding we see new characters like and
0
86,777
17,081,223,288
IssuesEvent
2021-07-08 05:37:54
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Overwrite the Startup Namer codelab with the content from Start Namer Null Safety
d: codelabs
The null unsafe [startup_namer](https://github.com/flutter/codelabs/tree/master/startup_namer) should be replaced with the content from [startup_namer_null_safety](https://github.com/flutter/codelabs/tree/master/startup_namer_null_safety).
1.0
Overwrite the Startup Namer codelab with the content from Start Namer Null Safety - The null unsafe [startup_namer](https://github.com/flutter/codelabs/tree/master/startup_namer) should be replaced with the content from [startup_namer_null_safety](https://github.com/flutter/codelabs/tree/master/startup_namer_null_safety).
non_priority
overwrite the startup namer codelab with the content from start namer null safety the null unsafe should be replaced with the content from
0
41,676
6,926,210,010
IssuesEvent
2017-11-30 18:22:03
CyberReboot/poseidon
https://api.github.com/repos/CyberReboot/poseidon
closed
limit info logging output
documentation enhancement
the reinvestigation loop outputs info logging every time through the loop. it would be easier to follow if you only did that output look in debug, and did info for when things in the investigation loop change state
1.0
limit info logging output - the reinvestigation loop outputs info logging every time through the loop. it would be easier to follow if you only did that output look in debug, and did info for when things in the investigation loop change state
non_priority
limit info logging output the reinvestigation loop outputs info logging every time through the loop it would be easier to follow if you only did that output look in debug and did info for when things in the investigation loop change state
0
234,317
25,829,991,302
IssuesEvent
2022-12-12 15:28:14
temporalio/samples-typescript
https://api.github.com/repos/temporalio/samples-typescript
closed
lerna-5.6.2.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lerna-5.6.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (lerna version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-3749](https://www.mend.io/vulnerability-database/CVE-2021-3749) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | axios-0.21.1.tgz | Transitive | 6.0.0 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3749</summary> ### Vulnerable Library - <b>axios-0.21.1.tgz</b></p> <p>Promise based HTTP client for the browser and node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.21.1.tgz">https://registry.npmjs.org/axios/-/axios-0.21.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/package.json</p> <p> Dependency Hierarchy: - lerna-5.6.2.tgz (Root Library) - nx-14.8.3.tgz - :x: **axios-0.21.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> axios is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3749>CVE-2021-3749</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/">https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (axios): 0.21.2</p> <p>Direct dependency fix Resolution (lerna): 6.0.0</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
lerna-5.6.2.tgz: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lerna-5.6.2.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (lerna version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-3749](https://www.mend.io/vulnerability-database/CVE-2021-3749) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | axios-0.21.1.tgz | Transitive | 6.0.0 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-3749</summary> ### Vulnerable Library - <b>axios-0.21.1.tgz</b></p> <p>Promise based HTTP client for the browser and node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.21.1.tgz">https://registry.npmjs.org/axios/-/axios-0.21.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/package.json</p> <p> Dependency Hierarchy: - lerna-5.6.2.tgz (Root Library) - nx-14.8.3.tgz - :x: **axios-0.21.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-typescript/commit/3b9bae980eb794fab08e5addef96e9ace5acb327">3b9bae980eb794fab08e5addef96e9ace5acb327</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> axios is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-08-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3749>CVE-2021-3749</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/">https://huntr.dev/bounties/1e8f07fc-c384-4ff9-8498-0690de2e8c31/</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (axios): 0.21.2</p> <p>Direct dependency fix Resolution (lerna): 6.0.0</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_priority
lerna tgz vulnerabilities highest severity is autoclosed vulnerable library lerna tgz path to dependency file package json path to vulnerable library node modules axios package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in lerna version remediation available high axios tgz transitive details cve vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file package json path to vulnerable library node modules axios package json dependency hierarchy lerna tgz root library nx tgz x axios tgz vulnerable library found in head commit a href found in base branch main vulnerability details axios is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution axios direct dependency fix resolution lerna rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
21,892
14,919,091,430
IssuesEvent
2021-01-22 23:08:15
aguirre-lab/ml4c3
https://api.github.com/repos/aguirre-lab/ml4c3
closed
New ml4c3 recipe: clustering
infrastructure 🚇
<!--- Provide a general summary of the issue in the Title above --> ## Why it Matters <!--- Describe why this new feature or enhancement should be implemented, etc --> ## Possible Solution <!-- High-level overview of how you propose to address --> ## Acceptance Criteria <!--- Unambiguous milestones; if any are incomplete, the PR cannot be merged --> ## Blocked by or Pending <!--- Links to other issues or PRs that block this issue and should be addressed or merged first. Also, label this issue with "`blocked 🚫`" -->
1.0
New ml4c3 recipe: clustering - <!--- Provide a general summary of the issue in the Title above --> ## Why it Matters <!--- Describe why this new feature or enhancement should be implemented, etc --> ## Possible Solution <!-- High-level overview of how you propose to address --> ## Acceptance Criteria <!--- Unambiguous milestones; if any are incomplete, the PR cannot be merged --> ## Blocked by or Pending <!--- Links to other issues or PRs that block this issue and should be addressed or merged first. Also, label this issue with "`blocked 🚫`" -->
non_priority
new recipe clustering why it matters possible solution acceptance criteria blocked by or pending
0
85,571
15,755,061,693
IssuesEvent
2021-03-31 01:06:52
LevyForchh/mapbox-gl-js
https://api.github.com/repos/LevyForchh/mapbox-gl-js
opened
CVE-2020-28500 (Medium) detected in lodash-4.17.15.tgz, lodash-4.17.11.tgz
security vulnerability
## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.15.tgz</b>, <b>lodash-4.17.11.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: mapbox-gl-js/package.json</p> <p>Path to vulnerable library: mapbox-gl-js/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - core-7.9.0.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: mapbox-gl-js/package.json</p> <p>Path to vulnerable library: mapbox-gl-js/node_modules/nyc/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - nyc-13.3.0.tgz (Root Library) - istanbul-reports-2.1.1.tgz - handlebars-4.1.0.tgz - async-2.6.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7">https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@babel/core:7.9.0;lodash:4.17.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"nyc:13.3.0;istanbul-reports:2.1.1;handlebars:4.1.0;async:2.6.2;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-28500 (Medium) detected in lodash-4.17.15.tgz, lodash-4.17.11.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.15.tgz</b>, <b>lodash-4.17.11.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: mapbox-gl-js/package.json</p> <p>Path to vulnerable library: mapbox-gl-js/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - core-7.9.0.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: mapbox-gl-js/package.json</p> <p>Path to vulnerable library: mapbox-gl-js/node_modules/nyc/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - nyc-13.3.0.tgz (Root Library) - istanbul-reports-2.1.1.tgz - handlebars-4.1.0.tgz - async-2.6.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7">https://github.com/lodash/lodash/commit/02906b8191d3c100c193fe6f7b27d1c40f200bb7</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.15","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@babel/core:7.9.0;lodash:4.17.15","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"nyc:13.3.0;istanbul-reports:2.1.1;handlebars:4.1.0;async:2.6.2;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-28500","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in lodash tgz lodash tgz cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file mapbox gl js package json path to vulnerable library mapbox gl js node modules lodash package json dependency hierarchy core tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file mapbox gl js package json path to vulnerable library mapbox gl js node modules nyc node modules lodash package json dependency hierarchy nyc tgz root library istanbul reports tgz handlebars tgz async tgz x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree babel core lodash isminimumfixversionavailable true minimumfixversion lodash packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency true dependencytree nyc istanbul reports handlebars async lodash isminimumfixversionavailable true minimumfixversion lodash basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions vulnerabilityurl
0
55,325
30,695,902,254
IssuesEvent
2023-07-26 18:34:26
lambdaclass/cairo-vm
https://api.github.com/repos/lambdaclass/cairo-vm
closed
Use simdjson to deserialize struct
enhancement performance blocked
Instead of using the serde library to deserialize programs, use the simdjson library https://github.com/simdjson/simdjson
True
Use simdjson to deserialize struct - Instead of using the serde library to deserialize programs, use the simdjson library https://github.com/simdjson/simdjson
non_priority
use simdjson to deserialize struct instead of using the serde library to deserialize programs use the simdjson library
0
200,149
15,791,093,591
IssuesEvent
2021-04-02 03:26:04
spacelab-ufsc/ttc2
https://api.github.com/repos/spacelab-ufsc/ttc2
opened
Documentation: Define a better nomenclature for the new mcu/radios architecture
documentation enhancement
Current names: - MCU main - MCU beacon - Main radio - Beacon radio
1.0
Documentation: Define a better nomenclature for the new mcu/radios architecture - Current names: - MCU main - MCU beacon - Main radio - Beacon radio
non_priority
documentation define a better nomenclature for the new mcu radios architecture current names mcu main mcu beacon main radio beacon radio
0
15,325
10,294,144,727
IssuesEvent
2019-08-27 19:03:30
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
QnA Maker follow-up prompts do not work in test WebChat
Pri3 cognitive-services/svc cxp needs-more-info triaged
I have been testing my KB in the QnA Maker Portal using the test chatbot and it doesn’t work as expected with follow-up prompts. The test chatbot respond partially as configured in the KB. However there isnt any of the follow-up prompts. Could you help and point me the right direction? Thank you very uch
1.0
QnA Maker follow-up prompts do not work in test WebChat - I have been testing my KB in the QnA Maker Portal using the test chatbot and it doesn’t work as expected with follow-up prompts. The test chatbot respond partially as configured in the KB. However there isnt any of the follow-up prompts. Could you help and point me the right direction? Thank you very uch
non_priority
qna maker follow up prompts do not work in test webchat i have been testing my kb in the qna maker portal using the test chatbot and it doesn’t work as expected with follow up prompts the test chatbot respond partially as configured in the kb however there isnt any of the follow up prompts could you help and point me the right direction thank you very uch
0
50,768
12,553,254,900
IssuesEvent
2020-06-06 21:14:12
spack/spack
https://api.github.com/repos/spack/spack
closed
Installation issue: [email protected]
build-error
<!-- Thanks for taking the time to report this build failure. To proceed with the report please: 1. Title the issue "Installation issue: <name-of-the-package>". 2. Provide the information required below. We encourage you to try, as much as possible, to reduce your problem to the minimal example that still reproduces the issue. That would help us a lot in fixing it quickly and effectively! --> ### Steps to reproduce the issue <!-- Fill in the exact spec you are trying to build and the relevant part of the error message --> From Ubuntu 20.04 with gcc 9.3.0 installed ```console $ spack install [email protected] ... In file included from /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.cc:190:0: /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_internal_defs.h:261:72: error: size of array 'assertion_failed__1150' is negative typedef char IMPL_PASTE(assertion_failed_##_, line)[2*(int)(pred)-1] ^ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_internal_defs.h:255:30: note: in expansion of macro 'IMPL_COMPILER_ASSERT' #define COMPILER_CHECK(pred) IMPL_COMPILER_ASSERT(pred, __LINE__) ^~~~~~~~~~~~~~~~~~~~ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.h:1461:3: note: in expansion of macro 'COMPILER_CHECK' COMPILER_CHECK(sizeof(((__sanitizer_##CLASS *) NULL)->MEMBER) == \ ^~~~~~~~~~~~~~ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.cc:1150:1: note: in expansion of macro 'CHECK_SIZE_AND_OFFSET' CHECK_SIZE_AND_OFFSET(ipc_perm, mode); ^~~~~~~~~~~~~~~~~~~~~ make[4]: *** [Makefile:523: sanitizer_platform_limits_posix.lo] Error 1 ``` ### Information on your system <!-- Please include the output of `spack debug report` --> ```console $ spack debug report * **Spack:** 0.14.2-1341-2ec680bbf * **Python:** 3.8.2 * **Platform:** linux-ubuntu20.04-x86_64 ``` <!-- If you have any relevant configuration detail (custom `packages.yaml` or `modules.yaml`, etc.) you can add that here as well. --> ### Additional information <!-- Please upload the following files. They should be present in the stage directory of the failing build. Also upload any config.log or similar file if one exists. --> * [spack-build-out.txt](https://github.com/spack/spack/files/4736493/spack-build-out.txt) * [spack-build-env.txt](https://github.com/spack/spack/files/4736502/spack-build-env.txt) <!-- Some packages have maintainers who have volunteered to debug build failures. Run `spack maintainers <name-of-the-package>` and @mention them here if they exist. --> gcc maintainer: @michaelkuhn ### General information <!-- These boxes can be checked by replacing [ ] with [x] or by clicking them after submitting the issue. --> - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [x] I have run `spack maintainers <name-of-the-package>` and @mentioned any maintainers - [x] I have uploaded the build log and environment files - [x] I have searched the issues of this repo and believe this is not a duplicate
1.0
Installation issue: [email protected] - <!-- Thanks for taking the time to report this build failure. To proceed with the report please: 1. Title the issue "Installation issue: <name-of-the-package>". 2. Provide the information required below. We encourage you to try, as much as possible, to reduce your problem to the minimal example that still reproduces the issue. That would help us a lot in fixing it quickly and effectively! --> ### Steps to reproduce the issue <!-- Fill in the exact spec you are trying to build and the relevant part of the error message --> From Ubuntu 20.04 with gcc 9.3.0 installed ```console $ spack install [email protected] ... In file included from /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.cc:190:0: /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_internal_defs.h:261:72: error: size of array 'assertion_failed__1150' is negative typedef char IMPL_PASTE(assertion_failed_##_, line)[2*(int)(pred)-1] ^ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_internal_defs.h:255:30: note: in expansion of macro 'IMPL_COMPILER_ASSERT' #define COMPILER_CHECK(pred) IMPL_COMPILER_ASSERT(pred, __LINE__) ^~~~~~~~~~~~~~~~~~~~ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.h:1461:3: note: in expansion of macro 'COMPILER_CHECK' COMPILER_CHECK(sizeof(((__sanitizer_##CLASS *) NULL)->MEMBER) == \ ^~~~~~~~~~~~~~ /tmp/jjolly/spack-stage/spack-stage-gcc-7.3.0-qew2co74raqn36lqlwarwxe7pn6vq63n/spack-src/libsanitizer/sanitizer_common/sanitizer_platform_limits_posix.cc:1150:1: note: in expansion of macro 'CHECK_SIZE_AND_OFFSET' CHECK_SIZE_AND_OFFSET(ipc_perm, mode); ^~~~~~~~~~~~~~~~~~~~~ make[4]: *** [Makefile:523: sanitizer_platform_limits_posix.lo] Error 1 ``` ### Information on your system <!-- Please include the output of `spack debug report` --> ```console $ spack debug report * **Spack:** 0.14.2-1341-2ec680bbf * **Python:** 3.8.2 * **Platform:** linux-ubuntu20.04-x86_64 ``` <!-- If you have any relevant configuration detail (custom `packages.yaml` or `modules.yaml`, etc.) you can add that here as well. --> ### Additional information <!-- Please upload the following files. They should be present in the stage directory of the failing build. Also upload any config.log or similar file if one exists. --> * [spack-build-out.txt](https://github.com/spack/spack/files/4736493/spack-build-out.txt) * [spack-build-env.txt](https://github.com/spack/spack/files/4736502/spack-build-env.txt) <!-- Some packages have maintainers who have volunteered to debug build failures. Run `spack maintainers <name-of-the-package>` and @mention them here if they exist. --> gcc maintainer: @michaelkuhn ### General information <!-- These boxes can be checked by replacing [ ] with [x] or by clicking them after submitting the issue. --> - [X] I have run `spack debug report` and reported the version of Spack/Python/Platform - [x] I have run `spack maintainers <name-of-the-package>` and @mentioned any maintainers - [x] I have uploaded the build log and environment files - [x] I have searched the issues of this repo and believe this is not a duplicate
non_priority
installation issue gcc thanks for taking the time to report this build failure to proceed with the report please title the issue installation issue provide the information required below we encourage you to try as much as possible to reduce your problem to the minimal example that still reproduces the issue that would help us a lot in fixing it quickly and effectively steps to reproduce the issue from ubuntu with gcc installed console spack install gcc in file included from tmp jjolly spack stage spack stage gcc spack src libsanitizer sanitizer common sanitizer platform limits posix cc tmp jjolly spack stage spack stage gcc spack src libsanitizer sanitizer common sanitizer internal defs h error size of array assertion failed is negative typedef char impl paste assertion failed line tmp jjolly spack stage spack stage gcc spack src libsanitizer sanitizer common sanitizer internal defs h note in expansion of macro impl compiler assert define compiler check pred impl compiler assert pred line tmp jjolly spack stage spack stage gcc spack src libsanitizer sanitizer common sanitizer platform limits posix h note in expansion of macro compiler check compiler check sizeof sanitizer class null member tmp jjolly spack stage spack stage gcc spack src libsanitizer sanitizer common sanitizer platform limits posix cc note in expansion of macro check size and offset check size and offset ipc perm mode make error information on your system console spack debug report spack python platform linux additional information and mention them here if they exist gcc maintainer michaelkuhn general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
88,710
25,490,522,082
IssuesEvent
2022-11-27 01:36:10
CleverRaven/Cataclysm-DDA
https://api.github.com/repos/CleverRaven/Cataclysm-DDA
closed
Cataclysm-DDA failed to build with "error C2015: too many characters in constant " with MSVC on Windows arm64ec
<Question> (S1 - Need confirmation) Code: Build OS: Windows
### Describe the bug Cataclysm-DDA failed to build with "F:\Cataclysm-DDA\src\unicode.cpp(105,17): error C2015: too many characters in constant " with MSVC on Windows arm64ec. It can reproduce on latest version on master branch. Could you please help look at this issue? **Error info:** ``` 10>F:\Cataclysm-DDA\src\unicode.cpp(105,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,62): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,71): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,63): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,62): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(110,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(110,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,27): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,36): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] ``` **Error log:** [Cataclysm-DDA_build_C2015.log](https://github.com/CleverRaven/Cataclysm-DDA/files/9741282/Cataclysm-DDA_build_C2015.log) **patch log** [Cataclysm-DDA_ARM64EC.patch.log](https://github.com/CleverRaven/Cataclysm-DDA/files/9741284/Cataclysm-DDA_ARM64EC.patch.log) ### Steps to reproduce 1. open x64 native tools command prompt for VS 2019 2. set VSCMD_SKIP_SENDTELEMETRY=1 & "C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\Common7\Tools\VsDevCmd.bat" -host_arch=amd64 -arch=arm64 3. git clone https://github.com/CleverRaven/Cataclysm-DDA F:\Cataclysm-DDA 4. git -C "F:\Cataclysm-DDA" apply f:\Cataclysm-DDA_ARM64EC.patch --ignore-space-change --ignore-whitespace --whitespace=nowarn 5. cd F:\vcpkg 6. bootstrap-vcpkg.bat 2>&1 7. vcpkg.exe install --recurse zlib sdl2 sdl2-ttf sdl2-image --triplet arm64-windows --clean-after-build 2>&1 8. F:\Cataclysm-DDA 9. mkdir build_arm64ec & cd build_arm64ec 10. cmake -G "Visual Studio 16 2019" -A ARM64EC -DCMAKE_SYSTEM_VERSION=10.0.22618.0 -DCMAKE_BUILD_TYPE=Release -DCURSES=OFF -DLOCALIZE=OFF -DTILES=1 -DBUILD_TESTING=OFF -DCMAKE_TOOLCHAIN_FILE=F:\vcpkg\scripts\buildsystems\vcpkg.cmake -DVCPKG_TARGET_TRIPLET=arm64-windows .. 11. set PATH=C:\tools\msys64\usr\bin;%PATH% 12. set _CL_=/permissive- 13. msbuild /m /p:Platform=ARM64EC /p:Configuration=Release CataclysmDDA.sln /t:Rebuild ### Expected behavior I expected build fine and does not error ### Screenshots _No response_ ### Versions and configuration OS: Windows Server 2022 Datacenter VS: Visual Studio 2019(16.11.19) ### Additional context _No response_
1.0
Cataclysm-DDA failed to build with "error C2015: too many characters in constant " with MSVC on Windows arm64ec - ### Describe the bug Cataclysm-DDA failed to build with "F:\Cataclysm-DDA\src\unicode.cpp(105,17): error C2015: too many characters in constant " with MSVC on Windows arm64ec. It can reproduce on latest version on master branch. Could you please help look at this issue? **Error info:** ``` 10>F:\Cataclysm-DDA\src\unicode.cpp(105,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,62): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(105,71): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(106,63): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(107,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(108,62): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,35): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,44): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(109,53): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(110,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(110,26): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,17): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,27): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] 10>F:\Cataclysm-DDA\src\unicode.cpp(111,36): error C2015: too many characters in constant [F:\Cataclysm-DDA\build_arm64ec\src\cataclysm-tiles-common.vcxproj] ``` **Error log:** [Cataclysm-DDA_build_C2015.log](https://github.com/CleverRaven/Cataclysm-DDA/files/9741282/Cataclysm-DDA_build_C2015.log) **patch log** [Cataclysm-DDA_ARM64EC.patch.log](https://github.com/CleverRaven/Cataclysm-DDA/files/9741284/Cataclysm-DDA_ARM64EC.patch.log) ### Steps to reproduce 1. open x64 native tools command prompt for VS 2019 2. set VSCMD_SKIP_SENDTELEMETRY=1 & "C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\Common7\Tools\VsDevCmd.bat" -host_arch=amd64 -arch=arm64 3. git clone https://github.com/CleverRaven/Cataclysm-DDA F:\Cataclysm-DDA 4. git -C "F:\Cataclysm-DDA" apply f:\Cataclysm-DDA_ARM64EC.patch --ignore-space-change --ignore-whitespace --whitespace=nowarn 5. cd F:\vcpkg 6. bootstrap-vcpkg.bat 2>&1 7. vcpkg.exe install --recurse zlib sdl2 sdl2-ttf sdl2-image --triplet arm64-windows --clean-after-build 2>&1 8. F:\Cataclysm-DDA 9. mkdir build_arm64ec & cd build_arm64ec 10. cmake -G "Visual Studio 16 2019" -A ARM64EC -DCMAKE_SYSTEM_VERSION=10.0.22618.0 -DCMAKE_BUILD_TYPE=Release -DCURSES=OFF -DLOCALIZE=OFF -DTILES=1 -DBUILD_TESTING=OFF -DCMAKE_TOOLCHAIN_FILE=F:\vcpkg\scripts\buildsystems\vcpkg.cmake -DVCPKG_TARGET_TRIPLET=arm64-windows .. 11. set PATH=C:\tools\msys64\usr\bin;%PATH% 12. set _CL_=/permissive- 13. msbuild /m /p:Platform=ARM64EC /p:Configuration=Release CataclysmDDA.sln /t:Rebuild ### Expected behavior I expected build fine and does not error ### Screenshots _No response_ ### Versions and configuration OS: Windows Server 2022 Datacenter VS: Visual Studio 2019(16.11.19) ### Additional context _No response_
non_priority
cataclysm dda failed to build with error too many characters in constant with msvc on windows describe the bug cataclysm dda failed to build with f cataclysm dda src unicode cpp error too many characters in constant with msvc on windows it can reproduce on latest version on master branch could you please help look at this issue error info f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant f cataclysm dda src unicode cpp error too many characters in constant error log patch log steps to reproduce open native tools command prompt for vs set vscmd skip sendtelemetry c program files microsoft visual studio enterprise tools vsdevcmd bat host arch arch git clone f cataclysm dda git c f cataclysm dda apply f cataclysm dda patch ignore space change ignore whitespace whitespace nowarn cd f vcpkg bootstrap vcpkg bat vcpkg exe install recurse zlib ttf image triplet windows clean after build f cataclysm dda mkdir build cd build cmake g visual studio a dcmake system version dcmake build type release dcurses off dlocalize off dtiles dbuild testing off dcmake toolchain file f vcpkg scripts buildsystems vcpkg cmake dvcpkg target triplet windows set path c tools usr bin path set cl permissive msbuild m p platform p configuration release cataclysmdda sln t rebuild expected behavior i expected build fine and does not error screenshots no response versions and configuration os windows server datacenter vs visual studio additional context no response
0
366,039
25,565,563,691
IssuesEvent
2022-11-30 14:04:25
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Update the reference docs to document the need to declare the native-maven-plugin when using buildpacks to create a native image
type: documentation
It should be mentioned in [this section](https://docs.spring.io/spring-boot/docs/3.0.0/reference/htmlsingle/#native-image.developing-your-first-application.buildpacks.maven), just as the equivalent Gradle plugin is mentioned in the [Gradle section](https://docs.spring.io/spring-boot/docs/3.0.0/reference/htmlsingle/#native-image.developing-your-first-application.buildpacks.gradle).
1.0
Update the reference docs to document the need to declare the native-maven-plugin when using buildpacks to create a native image - It should be mentioned in [this section](https://docs.spring.io/spring-boot/docs/3.0.0/reference/htmlsingle/#native-image.developing-your-first-application.buildpacks.maven), just as the equivalent Gradle plugin is mentioned in the [Gradle section](https://docs.spring.io/spring-boot/docs/3.0.0/reference/htmlsingle/#native-image.developing-your-first-application.buildpacks.gradle).
non_priority
update the reference docs to document the need to declare the native maven plugin when using buildpacks to create a native image it should be mentioned in just as the equivalent gradle plugin is mentioned in the
0
10,943
8,224,947,803
IssuesEvent
2018-09-06 14:57:49
cosmos/cosmos-sdk
https://api.github.com/repos/cosmos/cosmos-sdk
closed
`gaiacli keys add` UX could be much improved
API cli keys prelaunch security
`gaiacli keys add` is a bit weird. It has strange flags like: - `--no-backup`. Why would I ever run the command if people are watching? Seems unecessary - `--dry-run`. What's the point of this? Seems like we're just trying to generate a mnemonic and not make a key - that should probably be a separate command, otherwise we're just burdening this one with random flags - `--recover`. It's not always recovery. Sometimes we just want to make a new key from the same mnemonic (ie. with a different BIP44 path) - `--account` and `--index`. Not sure why this is better than just using a full. Let's just use the full path? - `--type` flag but currently only allow one type (secp256k1) I think we should leave `add` for legacy sake but add two new commands: `mnemonic` and `new`. ## Mnemonic ``` Generate a mnemonic from system or user-supplied entropy Usage: gaiacli keys mnemonic Flags: -h, --help help for add -u, --user bool Prompt the user to enter entropy. Otherwise, use the system's entropy ``` The output is just a straight up sequence of bip39 words to write down. By passing `--user` one can specify their own entropy (eg. by rolling ~99 die) instead of reading from system entropy. ## New ``` Usage: gaiacli keys new <name> [flags] Flags: --default Use system entropy to generate a new mnemonic and derive a key using default parameters -h, --help help for add --ledger Store a local reference to a private key on a Ledger device ``` This will completely replace `add`. Instead of flags, we use an interactive prompt that looks like: ``` $ gaiacli keys new bucky > Enter your bip39 mnemonic. > If you don't have one, just hit enter, and one will be generated for you. gloom bulb estate urge present page tissue volcano pause only ranch body eyebrow cruise island express describe reject road index radar marble release oyster > ------------------------------------- > Enter your bip39 passphrase. > If you don't have one, just hit enter, and the default "" will be used mybip39pass > ------------------------------------- > Enter your bip44 path. Default is 44'/0'/0'/0/0 44'/118'/0'/0/0 > ------------------------------------- > Enter password to encrypt the derived private key with. ondiskpass > ------------------------------------- Successfuly wrote encrypted priv key named "bucky" ``` The result is the creation of the key named `bucky` in the gaiacli db, just like `add` would have done. If you don't provide a mnemonic, one will be generated for you and output to the screen to be written down. To avoid the prompts, use `--default`. It will generate (and print) your mnemonic and use default values for the path and bip39 passphrase. It will still ask you for a password to encrypt your privkey on disk, of course. Related: - https://github.com/cosmos/cosmos-sdk/issues/2089
True
`gaiacli keys add` UX could be much improved - `gaiacli keys add` is a bit weird. It has strange flags like: - `--no-backup`. Why would I ever run the command if people are watching? Seems unecessary - `--dry-run`. What's the point of this? Seems like we're just trying to generate a mnemonic and not make a key - that should probably be a separate command, otherwise we're just burdening this one with random flags - `--recover`. It's not always recovery. Sometimes we just want to make a new key from the same mnemonic (ie. with a different BIP44 path) - `--account` and `--index`. Not sure why this is better than just using a full. Let's just use the full path? - `--type` flag but currently only allow one type (secp256k1) I think we should leave `add` for legacy sake but add two new commands: `mnemonic` and `new`. ## Mnemonic ``` Generate a mnemonic from system or user-supplied entropy Usage: gaiacli keys mnemonic Flags: -h, --help help for add -u, --user bool Prompt the user to enter entropy. Otherwise, use the system's entropy ``` The output is just a straight up sequence of bip39 words to write down. By passing `--user` one can specify their own entropy (eg. by rolling ~99 die) instead of reading from system entropy. ## New ``` Usage: gaiacli keys new <name> [flags] Flags: --default Use system entropy to generate a new mnemonic and derive a key using default parameters -h, --help help for add --ledger Store a local reference to a private key on a Ledger device ``` This will completely replace `add`. Instead of flags, we use an interactive prompt that looks like: ``` $ gaiacli keys new bucky > Enter your bip39 mnemonic. > If you don't have one, just hit enter, and one will be generated for you. gloom bulb estate urge present page tissue volcano pause only ranch body eyebrow cruise island express describe reject road index radar marble release oyster > ------------------------------------- > Enter your bip39 passphrase. > If you don't have one, just hit enter, and the default "" will be used mybip39pass > ------------------------------------- > Enter your bip44 path. Default is 44'/0'/0'/0/0 44'/118'/0'/0/0 > ------------------------------------- > Enter password to encrypt the derived private key with. ondiskpass > ------------------------------------- Successfuly wrote encrypted priv key named "bucky" ``` The result is the creation of the key named `bucky` in the gaiacli db, just like `add` would have done. If you don't provide a mnemonic, one will be generated for you and output to the screen to be written down. To avoid the prompts, use `--default`. It will generate (and print) your mnemonic and use default values for the path and bip39 passphrase. It will still ask you for a password to encrypt your privkey on disk, of course. Related: - https://github.com/cosmos/cosmos-sdk/issues/2089
non_priority
gaiacli keys add ux could be much improved gaiacli keys add is a bit weird it has strange flags like no backup why would i ever run the command if people are watching seems unecessary dry run what s the point of this seems like we re just trying to generate a mnemonic and not make a key that should probably be a separate command otherwise we re just burdening this one with random flags recover it s not always recovery sometimes we just want to make a new key from the same mnemonic ie with a different path account and index not sure why this is better than just using a full let s just use the full path type flag but currently only allow one type i think we should leave add for legacy sake but add two new commands mnemonic and new mnemonic generate a mnemonic from system or user supplied entropy usage gaiacli keys mnemonic flags h help help for add u user bool prompt the user to enter entropy otherwise use the system s entropy the output is just a straight up sequence of words to write down by passing user one can specify their own entropy eg by rolling die instead of reading from system entropy new usage gaiacli keys new flags default use system entropy to generate a new mnemonic and derive a key using default parameters h help help for add ledger store a local reference to a private key on a ledger device this will completely replace add instead of flags we use an interactive prompt that looks like gaiacli keys new bucky enter your mnemonic if you don t have one just hit enter and one will be generated for you gloom bulb estate urge present page tissue volcano pause only ranch body eyebrow cruise island express describe reject road index radar marble release oyster enter your passphrase if you don t have one just hit enter and the default will be used enter your path default is enter password to encrypt the derived private key with ondiskpass successfuly wrote encrypted priv key named bucky the result is the creation of the key named bucky in the gaiacli db just like add would have done if you don t provide a mnemonic one will be generated for you and output to the screen to be written down to avoid the prompts use default it will generate and print your mnemonic and use default values for the path and passphrase it will still ask you for a password to encrypt your privkey on disk of course related
0
377,979
26,277,751,421
IssuesEvent
2023-01-07 01:05:59
jonyboi396825/simplevectors
https://api.github.com/repos/jonyboi396825/simplevectors
closed
README
documentation
**Is there any incorrect information in the documentation? If so, describe.** The badges are on different lines and put them on the same line. Also add a period after "GNU make 3.81" in the beginning. **Is there anything that can be clarified/improved in the documentation? If so, describe.** Proofread
1.0
README - **Is there any incorrect information in the documentation? If so, describe.** The badges are on different lines and put them on the same line. Also add a period after "GNU make 3.81" in the beginning. **Is there anything that can be clarified/improved in the documentation? If so, describe.** Proofread
non_priority
readme is there any incorrect information in the documentation if so describe the badges are on different lines and put them on the same line also add a period after gnu make in the beginning is there anything that can be clarified improved in the documentation if so describe proofread
0
344,177
30,723,455,052
IssuesEvent
2023-07-27 17:40:39
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
JLM_Tests_class_SE80_0 FAILED: testThreadMXBeanProxy AssertionError: expected:<53> but was:<54>
test failure
Failure link ------------ https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_sanity.functional_x86-64_windows_Nightly_testList_0/568/consoleFull`win2019-x64-4a`: ``` 02:02:34 openjdk version "1.8.0_382-internal" 02:02:34 OpenJDK Runtime Environment (build 1.8.0_382-internal-****_2023_07_26_22_02-b00) 02:02:34 Eclipse OpenJ9 VM (build master-6140e2ff3fe, JRE 1.8.0 Windows Server 2019 amd64-64-Bit Compressed References 20230726_605 (JIT enabled, AOT enabled) 02:02:34 OpenJ9 - 6140e2ff3fe 02:02:34 OMR - 1dd4e595421 02:02:34 JCL - 535df28a5e2 based on jdk8u382-b04) ``` [Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&TARGET=testList+TESTLIST=JLM_Tests_class_SE80_0&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=Test_openjdk8_j9_sanity.functional_x86-64_windows_Nightly&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&VENDOR_TEST_DIRS=&EXTRA_DOCKER_ARGS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&PLATFORM=x86-64_windows&GENERATE_JOBS=true&KEEP_REPORTDIR=false&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&RERUN_ITERATIONS=0&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&LABEL=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=Nightly&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK8_x86-64_windows_Nightly%2F605%2FOpenJ9-JDK8-x86-64_windows-20230726-220225.tar.gz+https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK8_x86-64_windows_Nightly%2F605%2Ftest-images.tar.gz&JENKINS_KEY=&ADOPTOPENJDK_BRANCH=master&LIGHT_WEIGHT_CHECKOUT=true&USE_JRE=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&KEEP_WORKSPACE=false&USER_CREDENTIALS_ID=&JDK_VERSION=8&DOCKER_REGISTRY_URL=&ITERATIONS=1&VENDOR_TEST_REPOS=&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk8.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=6140e2ff3fe38897505dcee2a0ddde9ec6e82e03&JCK_GIT_REPO=&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&VENDOR_TEST_SHAS=&JDK_BRANCH=openj9&LABEL_ADDITION=&ARTIFACTORY_REPO=ci-openj9&ARTIFACTORY_ROOT_DIR=Test&UPSTREAM_TEST_JOB_NUMBER=570&DOCKERIMAGE_TAG=&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=&OPENJDK_SHA=&TRSS_URL=&USE_TESTENV_PROPERTIES=false&BUILD_LIST=functional&ADDITIONAL_ARTIFACTS_REQUIRED=&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&TIME_LIMIT=10&JVM_OPTIONS=&PARALLEL=None) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` 03:33:24 variation: NoOptions 03:33:25 JVM_OPTIONS: 03:33:35 FAILED: testThreadMXBeanProxy 03:33:35 java.lang.AssertionError: expected:<53> but was:<54> 03:33:35 at org.testng.AssertJUnit.fail(AssertJUnit.java:59) 03:33:35 at org.testng.AssertJUnit.failNotEquals(AssertJUnit.java:364) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:80) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:170) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:177) 03:33:35 at org.openj9.test.java.lang.management.TestManagementFactory.testThreadMXBeanProxy(TestManagementFactory.java:1316) 03:33:35 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 03:33:35 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 03:33:35 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 03:33:35 at java.lang.reflect.Method.invoke(Method.java:498) 03:33:35 at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) 03:33:35 at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) 03:33:35 at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) 03:33:35 at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) 03:33:35 at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) 03:33:35 at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) 03:33:35 at org.testng.TestRunner.privateRun(TestRunner.java:648) 03:33:35 at org.testng.TestRunner.run(TestRunner.java:505) 03:33:35 at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) 03:33:35 at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) 03:33:35 at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) 03:33:35 at org.testng.SuiteRunner.run(SuiteRunner.java:364) 03:33:35 at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) 03:33:35 at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) 03:33:35 at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) 03:33:35 at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) 03:33:35 at org.testng.TestNG.runSuites(TestNG.java:1049) 03:33:35 at org.testng.TestNG.run(TestNG.java:1017) 03:33:35 at org.testng.TestNG.privateMain(TestNG.java:1354) 03:33:35 at org.testng.TestNG.main(TestNG.java:1323) 03:33:35 03:33:35 03:33:35 =============================================== 03:33:35 JLM_Tests_class 03:33:35 Tests run: 66, Failures: 1, Skips: 0 03:33:35 =============================================== 03:33:35 03:33:35 03:33:35 =============================================== 03:33:35 JLM_Tests_Suite 03:33:35 Total tests run: 66, Failures: 1, Skips: 0 03:33:35 =============================================== 03:33:35 03:33:35 ----------------------------------- 03:33:35 JLM_Tests_class_SE80_0_FAILED ``` [50x grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2631/)
1.0
JLM_Tests_class_SE80_0 FAILED: testThreadMXBeanProxy AssertionError: expected:<53> but was:<54> - Failure link ------------ https://openj9-jenkins.osuosl.org/job/Test_openjdk8_j9_sanity.functional_x86-64_windows_Nightly_testList_0/568/consoleFull`win2019-x64-4a`: ``` 02:02:34 openjdk version "1.8.0_382-internal" 02:02:34 OpenJDK Runtime Environment (build 1.8.0_382-internal-****_2023_07_26_22_02-b00) 02:02:34 Eclipse OpenJ9 VM (build master-6140e2ff3fe, JRE 1.8.0 Windows Server 2019 amd64-64-Bit Compressed References 20230726_605 (JIT enabled, AOT enabled) 02:02:34 OpenJ9 - 6140e2ff3fe 02:02:34 OMR - 1dd4e595421 02:02:34 JCL - 535df28a5e2 based on jdk8u382-b04) ``` [Rerun in Grinder](https://openj9-jenkins.osuosl.org/job/Grinder/parambuild/?SDK_RESOURCE=customized&TARGET=testList+TESTLIST=JLM_Tests_class_SE80_0&TEST_FLAG=&UPSTREAM_TEST_JOB_NAME=Test_openjdk8_j9_sanity.functional_x86-64_windows_Nightly&DOCKER_REQUIRED=false&ACTIVE_NODE_TIMEOUT=&VENDOR_TEST_DIRS=&EXTRA_DOCKER_ARGS=&TKG_OWNER_BRANCH=adoptium%3Amaster&OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&PLATFORM=x86-64_windows&GENERATE_JOBS=true&KEEP_REPORTDIR=false&PERSONAL_BUILD=false&DOCKER_REGISTRY_DIR=&ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&RERUN_ITERATIONS=0&SETUP_JCK_RUN=false&DOCKER_REGISTRY_URL_CREDENTIAL_ID=&LABEL=&EXTRA_OPTIONS=&BUILD_IDENTIFIER=Nightly&CUSTOMIZED_SDK_URL=https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK8_x86-64_windows_Nightly%2F605%2FOpenJ9-JDK8-x86-64_windows-20230726-220225.tar.gz+https%3A%2F%2Fopenj9-artifactory.osuosl.org%2Fartifactory%2Fci-openj9%2FBuild_JDK8_x86-64_windows_Nightly%2F605%2Ftest-images.tar.gz&JENKINS_KEY=&ADOPTOPENJDK_BRANCH=master&LIGHT_WEIGHT_CHECKOUT=true&USE_JRE=false&ARTIFACTORY_SERVER=ci-eclipse-openj9&KEEP_WORKSPACE=false&USER_CREDENTIALS_ID=&JDK_VERSION=8&DOCKER_REGISTRY_URL=&ITERATIONS=1&VENDOR_TEST_REPOS=&JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk8.git&JCK_GIT_BRANCH=master&OPENJ9_BRANCH=master&OPENJ9_SHA=6140e2ff3fe38897505dcee2a0ddde9ec6e82e03&JCK_GIT_REPO=&VENDOR_TEST_BRANCHES=&OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&UPSTREAM_JOB_NAME=&CLOUD_PROVIDER=&CUSTOM_TARGET=&VENDOR_TEST_SHAS=&JDK_BRANCH=openj9&LABEL_ADDITION=&ARTIFACTORY_REPO=ci-openj9&ARTIFACTORY_ROOT_DIR=Test&UPSTREAM_TEST_JOB_NUMBER=570&DOCKERIMAGE_TAG=&JDK_IMPL=openj9&TEST_TIME=&SSH_AGENT_CREDENTIAL=&AUTO_DETECT=true&SLACK_CHANNEL=&DYNAMIC_COMPILE=true&RELATED_NODES=&ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&APPLICATION_OPTIONS=&CUSTOMIZED_SDK_URL_CREDENTIAL_ID=e047f50a-26c4-4b82-8525-6da17ec68450&ARCHIVE_TEST_RESULTS=false&NUM_MACHINES=&OPENJDK_SHA=&TRSS_URL=&USE_TESTENV_PROPERTIES=false&BUILD_LIST=functional&ADDITIONAL_ARTIFACTS_REQUIRED=&UPSTREAM_JOB_NUMBER=&STF_OWNER_BRANCH=adoptium%3Amaster&TIME_LIMIT=10&JVM_OPTIONS=&PARALLEL=None) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` 03:33:24 variation: NoOptions 03:33:25 JVM_OPTIONS: 03:33:35 FAILED: testThreadMXBeanProxy 03:33:35 java.lang.AssertionError: expected:<53> but was:<54> 03:33:35 at org.testng.AssertJUnit.fail(AssertJUnit.java:59) 03:33:35 at org.testng.AssertJUnit.failNotEquals(AssertJUnit.java:364) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:80) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:170) 03:33:35 at org.testng.AssertJUnit.assertEquals(AssertJUnit.java:177) 03:33:35 at org.openj9.test.java.lang.management.TestManagementFactory.testThreadMXBeanProxy(TestManagementFactory.java:1316) 03:33:35 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 03:33:35 at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 03:33:35 at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 03:33:35 at java.lang.reflect.Method.invoke(Method.java:498) 03:33:35 at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:124) 03:33:35 at org.testng.internal.Invoker.invokeMethod(Invoker.java:580) 03:33:35 at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:716) 03:33:35 at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:988) 03:33:35 at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:125) 03:33:35 at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:109) 03:33:35 at org.testng.TestRunner.privateRun(TestRunner.java:648) 03:33:35 at org.testng.TestRunner.run(TestRunner.java:505) 03:33:35 at org.testng.SuiteRunner.runTest(SuiteRunner.java:455) 03:33:35 at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:450) 03:33:35 at org.testng.SuiteRunner.privateRun(SuiteRunner.java:415) 03:33:35 at org.testng.SuiteRunner.run(SuiteRunner.java:364) 03:33:35 at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:52) 03:33:35 at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:84) 03:33:35 at org.testng.TestNG.runSuitesSequentially(TestNG.java:1208) 03:33:35 at org.testng.TestNG.runSuitesLocally(TestNG.java:1137) 03:33:35 at org.testng.TestNG.runSuites(TestNG.java:1049) 03:33:35 at org.testng.TestNG.run(TestNG.java:1017) 03:33:35 at org.testng.TestNG.privateMain(TestNG.java:1354) 03:33:35 at org.testng.TestNG.main(TestNG.java:1323) 03:33:35 03:33:35 03:33:35 =============================================== 03:33:35 JLM_Tests_class 03:33:35 Tests run: 66, Failures: 1, Skips: 0 03:33:35 =============================================== 03:33:35 03:33:35 03:33:35 =============================================== 03:33:35 JLM_Tests_Suite 03:33:35 Total tests run: 66, Failures: 1, Skips: 0 03:33:35 =============================================== 03:33:35 03:33:35 ----------------------------------- 03:33:35 JLM_Tests_class_SE80_0_FAILED ``` [50x grinder](https://openj9-jenkins.osuosl.org/job/Grinder/2631/)
non_priority
jlm tests class failed testthreadmxbeanproxy assertionerror expected but was failure link openjdk version internal openjdk runtime environment build internal eclipse vm build master jre windows server bit compressed references jit enabled aot enabled omr jcl based on change target to run only the failed test targets optional info failure output captured from console output variation nooptions jvm options failed testthreadmxbeanproxy java lang assertionerror expected but was at org testng assertjunit fail assertjunit java at org testng assertjunit failnotequals assertjunit java at org testng assertjunit assertequals assertjunit java at org testng assertjunit assertequals assertjunit java at org testng assertjunit assertequals assertjunit java at org test java lang management testmanagementfactory testthreadmxbeanproxy testmanagementfactory java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invoker invokemethod invoker java at org testng internal invoker invoketestmethod invoker java at org testng internal invoker invoketestmethods invoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at org testng testng privatemain testng java at org testng testng main testng java jlm tests class tests run failures skips jlm tests suite total tests run failures skips jlm tests class failed
0
28,025
13,495,725,807
IssuesEvent
2020-09-12 00:48:20
oneapi-src/oneDNN
https://api.github.com/repos/oneapi-src/oneDNN
closed
LSTM enhancement for batch size 1
enhancement performance
# Summary LSTM with single batch and multiple time steps doesn't work fast in oneDNN. # Problem statement My customer uses single batch with multiple sequence. It's usual in speech models. lstm_forward and lstm_prim.execute take some time from my analysis. FYI, one customer model includes ` batch size = 1 time steps = 63 input size = 24 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 31 input size = 48 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 15 input size = 96 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 31 input size = 48 channel size 32 gates = 4 Layer = 1 Direction 2 ` As shown above, the LSTMs have small feature(channel) size and input size. I tested the above cases using the below source code based on example: ``` void LSTM(memory::dim N, memory::dim T, memory::dim I, memory::dim C, memory::dim G, memory::dim L, memory::dim D) { using Clock = std::chrono::steady_clock; using OutputDur = std::chrono::microseconds; using std::chrono::duration_cast; // Create execution dnnl::engine. dnnl::engine engine(dnnl::engine::kind::cpu, 0); // Create dnnl::stream. dnnl::stream engine_stream(engine); // Source (src), weights, bias, and destination (dst) tensors // dimensions. memory::dims src_dims = {T, N, I}; memory::dims weights_dims = {L, D, I, G, C}; memory::dims weights_iter_dims = {L, D, C, G, C}; memory::dims bias_dims = {L, D, G, C}; memory::dims dst_dims = {T, N, D*C}; // Allocate buffers. std::vector<float> src_layer_data(product(src_dims)); std::vector<float> weights_layer_data(product(weights_dims)); std::vector<float> weights_iter_data(product(weights_iter_dims)); std::vector<float> dst_layer_data(product(dst_dims)); std::vector<float> bias_data(product(bias_dims)); // Initialize src, weights, and bias tensors. std::generate(src_layer_data.begin(), src_layer_data.end(), []() { static int i = 0; return std::cos(i++ / 10.f); }); std::generate(weights_layer_data.begin(), weights_layer_data.end(), []() { static int i = 0; return std::sin(i++ * 2.f); }); std::generate(bias_data.begin(), bias_data.end(), []() { static int i = 0; return std::tanh(i++); }); // Create memory descriptors and memory objects for src, bias, and dst. auto src_layer_md = memory::desc(src_dims, dt::f32, tag::tnc); auto bias_md = memory::desc(bias_dims, dt::f32, tag::ldgo); auto dst_layer_md = memory::desc(dst_dims, dt::f32, tag::tnc); auto src_layer_mem = memory(src_layer_md, engine); auto bias_mem = memory(bias_md, engine); auto dst_layer_mem = memory(dst_layer_md, engine); // Create memory objects for weights using user's memory layout. In this // example, LDIGO is assumed. auto user_weights_layer_mem = memory({weights_dims, dt::f32, tag::ldigo}, engine); auto user_weights_iter_mem = memory({weights_iter_dims, dt::f32, tag::ldigo}, engine); // Write data to memory object's handle. write_to_dnnl_memory(src_layer_data.data(), src_layer_mem); write_to_dnnl_memory(bias_data.data(), bias_mem); write_to_dnnl_memory(weights_layer_data.data(), user_weights_layer_mem); write_to_dnnl_memory(weights_iter_data.data(), user_weights_iter_mem); // Create memory descriptors for weights with format_tag::any. This enables // the LSTM primitive to choose the optimized memory layout. auto lstm_weights_layer_md = memory::desc(weights_dims, dt::f32, tag::any); auto lstm_weights_iter_md = memory::desc(weights_iter_dims, dt::f32, tag::any); // Optional memory descriptors for recurrent data. auto src_iter_md = memory::desc(); auto src_iter_c_md = memory::desc(); auto dst_iter_md = memory::desc(); auto dst_iter_c_md = memory::desc(); // Create operation descriptor. auto lstm_desc = lstm_forward::desc(prop_kind::forward_training, D > 1 ? rnn_direction::bidirectional_concat : rnn_direction::unidirectional_left2right, src_layer_md, src_iter_md, src_iter_c_md, lstm_weights_layer_md, lstm_weights_iter_md, bias_md, dst_layer_md, dst_iter_md, dst_iter_c_md); // Create primitive descriptor. auto lstm_pd = lstm_forward::primitive_desc(lstm_desc, engine); // For now, assume that the weights memory layout generated by the primitive // and the ones provided by the user are identical. auto lstm_weights_layer_mem = user_weights_layer_mem; auto lstm_weights_iter_mem = user_weights_iter_mem; // Reorder the data in case the weights memory layout generated by the // primitive and the one provided by the user are different. In this case, // we create additional memory objects with internal buffers that will // contain the reordered data. if (lstm_pd.weights_desc() != user_weights_layer_mem.get_desc()) { lstm_weights_layer_mem = memory(lstm_pd.weights_desc(), engine); reorder(user_weights_layer_mem, lstm_weights_layer_mem) .execute(engine_stream, user_weights_layer_mem, lstm_weights_layer_mem); } if (lstm_pd.weights_iter_desc() != user_weights_iter_mem.get_desc()) { lstm_weights_iter_mem = memory(lstm_pd.weights_iter_desc(), engine); reorder(user_weights_iter_mem, lstm_weights_iter_mem) .execute(engine_stream, user_weights_iter_mem, lstm_weights_iter_mem); } // Create the memory objects from the primitive descriptor. A workspace is // also required for LSTM. // NOTE: Here, the workspace is required for later usage in backward // propagation mode. auto src_iter_mem = memory(lstm_pd.src_iter_desc(), engine); auto src_iter_c_mem = memory(lstm_pd.src_iter_c_desc(), engine); auto weights_iter_mem = memory(lstm_pd.weights_iter_desc(), engine); auto dst_iter_mem = memory(lstm_pd.dst_iter_desc(), engine); auto dst_iter_c_mem = memory(lstm_pd.dst_iter_c_desc(), engine); auto workspace_mem = memory(lstm_pd.workspace_desc(), engine); // Create the primitive. const auto pdf_time1 = Clock::now(); auto lstm_prim = lstm_forward(lstm_pd); const auto pdf_time2 = Clock::now(); // Primitive arguments std::unordered_map<int, memory> lstm_args; lstm_args.insert({DNNL_ARG_SRC_LAYER, src_layer_mem}); lstm_args.insert({DNNL_ARG_WEIGHTS_LAYER, lstm_weights_layer_mem}); lstm_args.insert({DNNL_ARG_WEIGHTS_ITER, lstm_weights_iter_mem}); lstm_args.insert({DNNL_ARG_BIAS, bias_mem}); lstm_args.insert({DNNL_ARG_DST_LAYER, dst_layer_mem}); lstm_args.insert({DNNL_ARG_SRC_ITER, src_iter_mem}); lstm_args.insert({DNNL_ARG_SRC_ITER_C, src_iter_c_mem}); lstm_args.insert({DNNL_ARG_DST_ITER, dst_iter_mem}); lstm_args.insert({DNNL_ARG_DST_ITER_C, dst_iter_c_mem}); lstm_args.insert({DNNL_ARG_WORKSPACE, workspace_mem}); // Primitive execution: LSTM. const auto exe_time1 = Clock::now(); lstm_prim.execute(engine_stream, lstm_args); // Wait for the computation to finalize. engine_stream.wait(); const auto exe_time2 = Clock::now(); const auto pdf_time = pdf_time2-pdf_time1; const auto exe_time = exe_time2-exe_time1; std::ofstream f; f.open ("/home/oneDNN/workspace/lstm_result.txt",std::ios::app); f << "[Batch: " << N << "] [Seq: " << T << "] [InputSize: " << I << "] [FeatureSize: " << C << "] [Directions: " << D << " ] " << duration_cast<OutputDur>(pdf_time).count() << ", " << duration_cast<OutputDur>(exe_time).count() << "\n"; myfile.close(); } ``` Results(microseconds): [Batch: 1] [Seq: 63] [InputSize: 24] [FeatureSize: 32] [Directions: 2 ] 17, 1075 [Batch: 1] [Seq: 31] [InputSize: 48] [FeatureSize: 32] [Directions: 2 ] 166, 595 [Batch: 1] [Seq: 15] [InputSize: 96] [FeatureSize: 32] [Directions: 2 ] 187, 348 Verbose echo: dnnl_verbose,info,oneDNN v1.6.0 (commit N/A) dnnl_verbose,info,cpu,runtime:TBB dnnl_verbose,info,cpu,isa:Intel AVX-512 with AVX512BW, AVX512VL, and AVX512DQ extensions dnnl_verbose,info,gpu,runtime:none dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t63mb1sic32slc24dhc32dlc64,1.05591 dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t31mb1sic32slc48dhc32dlc64,0.533936 dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t15mb1sic32slc96dhc32dlc64,0.3059081 # Preferred solution Would be nice if oneDNN provides well-optimized LSTM for single batch case.
True
LSTM enhancement for batch size 1 - # Summary LSTM with single batch and multiple time steps doesn't work fast in oneDNN. # Problem statement My customer uses single batch with multiple sequence. It's usual in speech models. lstm_forward and lstm_prim.execute take some time from my analysis. FYI, one customer model includes ` batch size = 1 time steps = 63 input size = 24 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 31 input size = 48 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 15 input size = 96 channel size 32 gates = 4 Layer = 1 Direction 2 ` ` batch size = 1 time steps = 31 input size = 48 channel size 32 gates = 4 Layer = 1 Direction 2 ` As shown above, the LSTMs have small feature(channel) size and input size. I tested the above cases using the below source code based on example: ``` void LSTM(memory::dim N, memory::dim T, memory::dim I, memory::dim C, memory::dim G, memory::dim L, memory::dim D) { using Clock = std::chrono::steady_clock; using OutputDur = std::chrono::microseconds; using std::chrono::duration_cast; // Create execution dnnl::engine. dnnl::engine engine(dnnl::engine::kind::cpu, 0); // Create dnnl::stream. dnnl::stream engine_stream(engine); // Source (src), weights, bias, and destination (dst) tensors // dimensions. memory::dims src_dims = {T, N, I}; memory::dims weights_dims = {L, D, I, G, C}; memory::dims weights_iter_dims = {L, D, C, G, C}; memory::dims bias_dims = {L, D, G, C}; memory::dims dst_dims = {T, N, D*C}; // Allocate buffers. std::vector<float> src_layer_data(product(src_dims)); std::vector<float> weights_layer_data(product(weights_dims)); std::vector<float> weights_iter_data(product(weights_iter_dims)); std::vector<float> dst_layer_data(product(dst_dims)); std::vector<float> bias_data(product(bias_dims)); // Initialize src, weights, and bias tensors. std::generate(src_layer_data.begin(), src_layer_data.end(), []() { static int i = 0; return std::cos(i++ / 10.f); }); std::generate(weights_layer_data.begin(), weights_layer_data.end(), []() { static int i = 0; return std::sin(i++ * 2.f); }); std::generate(bias_data.begin(), bias_data.end(), []() { static int i = 0; return std::tanh(i++); }); // Create memory descriptors and memory objects for src, bias, and dst. auto src_layer_md = memory::desc(src_dims, dt::f32, tag::tnc); auto bias_md = memory::desc(bias_dims, dt::f32, tag::ldgo); auto dst_layer_md = memory::desc(dst_dims, dt::f32, tag::tnc); auto src_layer_mem = memory(src_layer_md, engine); auto bias_mem = memory(bias_md, engine); auto dst_layer_mem = memory(dst_layer_md, engine); // Create memory objects for weights using user's memory layout. In this // example, LDIGO is assumed. auto user_weights_layer_mem = memory({weights_dims, dt::f32, tag::ldigo}, engine); auto user_weights_iter_mem = memory({weights_iter_dims, dt::f32, tag::ldigo}, engine); // Write data to memory object's handle. write_to_dnnl_memory(src_layer_data.data(), src_layer_mem); write_to_dnnl_memory(bias_data.data(), bias_mem); write_to_dnnl_memory(weights_layer_data.data(), user_weights_layer_mem); write_to_dnnl_memory(weights_iter_data.data(), user_weights_iter_mem); // Create memory descriptors for weights with format_tag::any. This enables // the LSTM primitive to choose the optimized memory layout. auto lstm_weights_layer_md = memory::desc(weights_dims, dt::f32, tag::any); auto lstm_weights_iter_md = memory::desc(weights_iter_dims, dt::f32, tag::any); // Optional memory descriptors for recurrent data. auto src_iter_md = memory::desc(); auto src_iter_c_md = memory::desc(); auto dst_iter_md = memory::desc(); auto dst_iter_c_md = memory::desc(); // Create operation descriptor. auto lstm_desc = lstm_forward::desc(prop_kind::forward_training, D > 1 ? rnn_direction::bidirectional_concat : rnn_direction::unidirectional_left2right, src_layer_md, src_iter_md, src_iter_c_md, lstm_weights_layer_md, lstm_weights_iter_md, bias_md, dst_layer_md, dst_iter_md, dst_iter_c_md); // Create primitive descriptor. auto lstm_pd = lstm_forward::primitive_desc(lstm_desc, engine); // For now, assume that the weights memory layout generated by the primitive // and the ones provided by the user are identical. auto lstm_weights_layer_mem = user_weights_layer_mem; auto lstm_weights_iter_mem = user_weights_iter_mem; // Reorder the data in case the weights memory layout generated by the // primitive and the one provided by the user are different. In this case, // we create additional memory objects with internal buffers that will // contain the reordered data. if (lstm_pd.weights_desc() != user_weights_layer_mem.get_desc()) { lstm_weights_layer_mem = memory(lstm_pd.weights_desc(), engine); reorder(user_weights_layer_mem, lstm_weights_layer_mem) .execute(engine_stream, user_weights_layer_mem, lstm_weights_layer_mem); } if (lstm_pd.weights_iter_desc() != user_weights_iter_mem.get_desc()) { lstm_weights_iter_mem = memory(lstm_pd.weights_iter_desc(), engine); reorder(user_weights_iter_mem, lstm_weights_iter_mem) .execute(engine_stream, user_weights_iter_mem, lstm_weights_iter_mem); } // Create the memory objects from the primitive descriptor. A workspace is // also required for LSTM. // NOTE: Here, the workspace is required for later usage in backward // propagation mode. auto src_iter_mem = memory(lstm_pd.src_iter_desc(), engine); auto src_iter_c_mem = memory(lstm_pd.src_iter_c_desc(), engine); auto weights_iter_mem = memory(lstm_pd.weights_iter_desc(), engine); auto dst_iter_mem = memory(lstm_pd.dst_iter_desc(), engine); auto dst_iter_c_mem = memory(lstm_pd.dst_iter_c_desc(), engine); auto workspace_mem = memory(lstm_pd.workspace_desc(), engine); // Create the primitive. const auto pdf_time1 = Clock::now(); auto lstm_prim = lstm_forward(lstm_pd); const auto pdf_time2 = Clock::now(); // Primitive arguments std::unordered_map<int, memory> lstm_args; lstm_args.insert({DNNL_ARG_SRC_LAYER, src_layer_mem}); lstm_args.insert({DNNL_ARG_WEIGHTS_LAYER, lstm_weights_layer_mem}); lstm_args.insert({DNNL_ARG_WEIGHTS_ITER, lstm_weights_iter_mem}); lstm_args.insert({DNNL_ARG_BIAS, bias_mem}); lstm_args.insert({DNNL_ARG_DST_LAYER, dst_layer_mem}); lstm_args.insert({DNNL_ARG_SRC_ITER, src_iter_mem}); lstm_args.insert({DNNL_ARG_SRC_ITER_C, src_iter_c_mem}); lstm_args.insert({DNNL_ARG_DST_ITER, dst_iter_mem}); lstm_args.insert({DNNL_ARG_DST_ITER_C, dst_iter_c_mem}); lstm_args.insert({DNNL_ARG_WORKSPACE, workspace_mem}); // Primitive execution: LSTM. const auto exe_time1 = Clock::now(); lstm_prim.execute(engine_stream, lstm_args); // Wait for the computation to finalize. engine_stream.wait(); const auto exe_time2 = Clock::now(); const auto pdf_time = pdf_time2-pdf_time1; const auto exe_time = exe_time2-exe_time1; std::ofstream f; f.open ("/home/oneDNN/workspace/lstm_result.txt",std::ios::app); f << "[Batch: " << N << "] [Seq: " << T << "] [InputSize: " << I << "] [FeatureSize: " << C << "] [Directions: " << D << " ] " << duration_cast<OutputDur>(pdf_time).count() << ", " << duration_cast<OutputDur>(exe_time).count() << "\n"; myfile.close(); } ``` Results(microseconds): [Batch: 1] [Seq: 63] [InputSize: 24] [FeatureSize: 32] [Directions: 2 ] 17, 1075 [Batch: 1] [Seq: 31] [InputSize: 48] [FeatureSize: 32] [Directions: 2 ] 166, 595 [Batch: 1] [Seq: 15] [InputSize: 96] [FeatureSize: 32] [Directions: 2 ] 187, 348 Verbose echo: dnnl_verbose,info,oneDNN v1.6.0 (commit N/A) dnnl_verbose,info,cpu,runtime:TBB dnnl_verbose,info,cpu,isa:Intel AVX-512 with AVX512BW, AVX512VL, and AVX512DQ extensions dnnl_verbose,info,gpu,runtime:none dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t63mb1sic32slc24dhc32dlc64,1.05591 dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t31mb1sic32slc48dhc32dlc64,0.533936 dnnl_verbose,exec,cpu,rnn,ref:any,forward_training,src_layer_f32::blocked:abc:f0 src_iter_undef::undef::f0 wei_layer_f32::blocked:abcde:f0 wei_iter_f32::blocked:abcde:f0 bias_f32::blocked:abcd:f0 dst_layer_f32::blocked:abc:f0 dst_iter_undef::undef::f0,,alg:vanilla_lstm direction:bidirectional_concat activation:undef,l1t15mb1sic32slc96dhc32dlc64,0.3059081 # Preferred solution Would be nice if oneDNN provides well-optimized LSTM for single batch case.
non_priority
lstm enhancement for batch size summary lstm with single batch and multiple time steps doesn t work fast in onednn problem statement my customer uses single batch with multiple sequence it s usual in speech models lstm forward and lstm prim execute take some time from my analysis fyi one customer model includes batch size time steps input size channel size gates layer direction batch size time steps input size channel size gates layer direction batch size time steps input size channel size gates layer direction batch size time steps input size channel size gates layer direction as shown above the lstms have small feature channel size and input size i tested the above cases using the below source code based on example void lstm memory dim n memory dim t memory dim i memory dim c memory dim g memory dim l memory dim d using clock std chrono steady clock using outputdur std chrono microseconds using std chrono duration cast create execution dnnl engine dnnl engine engine dnnl engine kind cpu create dnnl stream dnnl stream engine stream engine source src weights bias and destination dst tensors dimensions memory dims src dims t n i memory dims weights dims l d i g c memory dims weights iter dims l d c g c memory dims bias dims l d g c memory dims dst dims t n d c allocate buffers std vector src layer data product src dims std vector weights layer data product weights dims std vector weights iter data product weights iter dims std vector dst layer data product dst dims std vector bias data product bias dims initialize src weights and bias tensors std generate src layer data begin src layer data end static int i return std cos i f std generate weights layer data begin weights layer data end static int i return std sin i f std generate bias data begin bias data end static int i return std tanh i create memory descriptors and memory objects for src bias and dst auto src layer md memory desc src dims dt tag tnc auto bias md memory desc bias dims dt tag ldgo auto dst layer md memory desc dst dims dt tag tnc auto src layer mem memory src layer md engine auto bias mem memory bias md engine auto dst layer mem memory dst layer md engine create memory objects for weights using user s memory layout in this example ldigo is assumed auto user weights layer mem memory weights dims dt tag ldigo engine auto user weights iter mem memory weights iter dims dt tag ldigo engine write data to memory object s handle write to dnnl memory src layer data data src layer mem write to dnnl memory bias data data bias mem write to dnnl memory weights layer data data user weights layer mem write to dnnl memory weights iter data data user weights iter mem create memory descriptors for weights with format tag any this enables the lstm primitive to choose the optimized memory layout auto lstm weights layer md memory desc weights dims dt tag any auto lstm weights iter md memory desc weights iter dims dt tag any optional memory descriptors for recurrent data auto src iter md memory desc auto src iter c md memory desc auto dst iter md memory desc auto dst iter c md memory desc create operation descriptor auto lstm desc lstm forward desc prop kind forward training d rnn direction bidirectional concat rnn direction unidirectional src layer md src iter md src iter c md lstm weights layer md lstm weights iter md bias md dst layer md dst iter md dst iter c md create primitive descriptor auto lstm pd lstm forward primitive desc lstm desc engine for now assume that the weights memory layout generated by the primitive and the ones provided by the user are identical auto lstm weights layer mem user weights layer mem auto lstm weights iter mem user weights iter mem reorder the data in case the weights memory layout generated by the primitive and the one provided by the user are different in this case we create additional memory objects with internal buffers that will contain the reordered data if lstm pd weights desc user weights layer mem get desc lstm weights layer mem memory lstm pd weights desc engine reorder user weights layer mem lstm weights layer mem execute engine stream user weights layer mem lstm weights layer mem if lstm pd weights iter desc user weights iter mem get desc lstm weights iter mem memory lstm pd weights iter desc engine reorder user weights iter mem lstm weights iter mem execute engine stream user weights iter mem lstm weights iter mem create the memory objects from the primitive descriptor a workspace is also required for lstm note here the workspace is required for later usage in backward propagation mode auto src iter mem memory lstm pd src iter desc engine auto src iter c mem memory lstm pd src iter c desc engine auto weights iter mem memory lstm pd weights iter desc engine auto dst iter mem memory lstm pd dst iter desc engine auto dst iter c mem memory lstm pd dst iter c desc engine auto workspace mem memory lstm pd workspace desc engine create the primitive const auto pdf clock now auto lstm prim lstm forward lstm pd const auto pdf clock now primitive arguments std unordered map lstm args lstm args insert dnnl arg src layer src layer mem lstm args insert dnnl arg weights layer lstm weights layer mem lstm args insert dnnl arg weights iter lstm weights iter mem lstm args insert dnnl arg bias bias mem lstm args insert dnnl arg dst layer dst layer mem lstm args insert dnnl arg src iter src iter mem lstm args insert dnnl arg src iter c src iter c mem lstm args insert dnnl arg dst iter dst iter mem lstm args insert dnnl arg dst iter c dst iter c mem lstm args insert dnnl arg workspace workspace mem primitive execution lstm const auto exe clock now lstm prim execute engine stream lstm args wait for the computation to finalize engine stream wait const auto exe clock now const auto pdf time pdf pdf const auto exe time exe exe std ofstream f f open home onednn workspace lstm result txt std ios app f pdf time count exe time count n myfile close results microseconds verbose echo dnnl verbose info onednn commit n a dnnl verbose info cpu runtime tbb dnnl verbose info cpu isa intel avx with and extensions dnnl verbose info gpu runtime none dnnl verbose exec cpu rnn ref any forward training src layer blocked abc src iter undef undef wei layer blocked abcde wei iter blocked abcde bias blocked abcd dst layer blocked abc dst iter undef undef alg vanilla lstm direction bidirectional concat activation undef dnnl verbose exec cpu rnn ref any forward training src layer blocked abc src iter undef undef wei layer blocked abcde wei iter blocked abcde bias blocked abcd dst layer blocked abc dst iter undef undef alg vanilla lstm direction bidirectional concat activation undef dnnl verbose exec cpu rnn ref any forward training src layer blocked abc src iter undef undef wei layer blocked abcde wei iter blocked abcde bias blocked abcd dst layer blocked abc dst iter undef undef alg vanilla lstm direction bidirectional concat activation undef preferred solution would be nice if onednn provides well optimized lstm for single batch case
0
31,890
13,652,811,615
IssuesEvent
2020-09-27 09:29:26
tuna/issues
https://api.github.com/repos/tuna/issues
closed
[tuna]404 at /anaconda/cloud/pytorch/win-64/current_repodata.json
Service Issue
<!-- 请使用此模板来报告 bug,并尽可能多地提供信息。 Please use this template while reporting a bug and provide as much info as possible. --> conda install django-request Collecting package metadata (current_repodata.json): failed #### 期望的现象(What you expected to happen) #### 如何重现(How to reproduce it) #### 其他事项(Anything else we need to know) #### 您的环境(Environment) - 操作系统(OS Version): - 浏览器(如果适用)(Browser version, if applicable): - 其他(Others):
1.0
[tuna]404 at /anaconda/cloud/pytorch/win-64/current_repodata.json - <!-- 请使用此模板来报告 bug,并尽可能多地提供信息。 Please use this template while reporting a bug and provide as much info as possible. --> conda install django-request Collecting package metadata (current_repodata.json): failed #### 期望的现象(What you expected to happen) #### 如何重现(How to reproduce it) #### 其他事项(Anything else we need to know) #### 您的环境(Environment) - 操作系统(OS Version): - 浏览器(如果适用)(Browser version, if applicable): - 其他(Others):
non_priority
at anaconda cloud pytorch win current repodata json 请使用此模板来报告 bug,并尽可能多地提供信息。 please use this template while reporting a bug and provide as much info as possible conda install django request collecting package metadata current repodata json failed 期望的现象(what you expected to happen) 如何重现(how to reproduce it) 其他事项(anything else we need to know) 您的环境(environment) 操作系统(os version): 浏览器(如果适用)(browser version if applicable): 其他(others):
0
23,994
11,996,482,243
IssuesEvent
2020-04-08 16:50:16
microsoft/botframework-sdk
https://api.github.com/repos/microsoft/botframework-sdk
closed
bot returning welcome message again when i click on button
Bot Services customer-reported
```node // Copyright (c) Microsoft Corporation. All rights reserved. // Licensed under the MIT License. const { CardFactory, ActionTypes, ActivityHandler, MessageFactory } = require('botbuilder'); const { DialogBot } = require('./dialogBot'); var BotResponses = require('../constant/string'); var ConstantVariables = require('../constant/constantValue'); class DialogAndWelcomeBot extends DialogBot { constructor(conversationState, userState, dialog) { super(conversationState, userState, dialog); this.onMembersAdded(async (context, next) => { const membersAdded = context.activity.membersAdded; for (let cnt = 0; cnt < membersAdded.length; cnt++) { if (membersAdded[cnt].id !== context.activity.recipient.id) { await context.sendActivity(BotResponses.GREET); await context.sendActivity(BotResponses.GREET1); const buttons = [ { type: ActionTypes.ImBack, title: ConstantVariables.SELF_ASSESSMENT, value: ConstantVariables.SELF_ASSESSMENT }, { type: ActionTypes.ImBack, title: ConstantVariables.sympton, value: ConstantVariables.sympton } ]; const card = CardFactory.heroCard('', undefined, buttons, { text: 'Please select one of the categories below or type in your question.' }); await context.sendActivity({ attachments: [card] }); // await dialog.run(context, conversationState.createProperty('DialogState')); } } // By calling next() you ensure that the next BotHandler is run. await next(); }); } } module.exports.DialogAndWelcomeBot = DialogAndWelcomeBot; ```
1.0
bot returning welcome message again when i click on button - ```node // Copyright (c) Microsoft Corporation. All rights reserved. // Licensed under the MIT License. const { CardFactory, ActionTypes, ActivityHandler, MessageFactory } = require('botbuilder'); const { DialogBot } = require('./dialogBot'); var BotResponses = require('../constant/string'); var ConstantVariables = require('../constant/constantValue'); class DialogAndWelcomeBot extends DialogBot { constructor(conversationState, userState, dialog) { super(conversationState, userState, dialog); this.onMembersAdded(async (context, next) => { const membersAdded = context.activity.membersAdded; for (let cnt = 0; cnt < membersAdded.length; cnt++) { if (membersAdded[cnt].id !== context.activity.recipient.id) { await context.sendActivity(BotResponses.GREET); await context.sendActivity(BotResponses.GREET1); const buttons = [ { type: ActionTypes.ImBack, title: ConstantVariables.SELF_ASSESSMENT, value: ConstantVariables.SELF_ASSESSMENT }, { type: ActionTypes.ImBack, title: ConstantVariables.sympton, value: ConstantVariables.sympton } ]; const card = CardFactory.heroCard('', undefined, buttons, { text: 'Please select one of the categories below or type in your question.' }); await context.sendActivity({ attachments: [card] }); // await dialog.run(context, conversationState.createProperty('DialogState')); } } // By calling next() you ensure that the next BotHandler is run. await next(); }); } } module.exports.DialogAndWelcomeBot = DialogAndWelcomeBot; ```
non_priority
bot returning welcome message again when i click on button node copyright c microsoft corporation all rights reserved licensed under the mit license const cardfactory actiontypes activityhandler messagefactory require botbuilder const dialogbot require dialogbot var botresponses require constant string var constantvariables require constant constantvalue class dialogandwelcomebot extends dialogbot constructor conversationstate userstate dialog super conversationstate userstate dialog this onmembersadded async context next const membersadded context activity membersadded for let cnt cnt membersadded length cnt if membersadded id context activity recipient id await context sendactivity botresponses greet await context sendactivity botresponses const buttons type actiontypes imback title constantvariables self assessment value constantvariables self assessment type actiontypes imback title constantvariables sympton value constantvariables sympton const card cardfactory herocard undefined buttons text please select one of the categories below or type in your question await context sendactivity attachments await dialog run context conversationstate createproperty dialogstate by calling next you ensure that the next bothandler is run await next module exports dialogandwelcomebot dialogandwelcomebot
0
27,779
8,035,333,367
IssuesEvent
2018-07-30 03:46:27
SoftEtherVPN/SoftEtherVPN
https://api.github.com/repos/SoftEtherVPN/SoftEtherVPN
opened
Building with the MinGW toolchain
build & release feature
https://github.com/SoftEtherVPN/SoftEtherVPN/issues/65#issuecomment-373994100 > I'd also like to know how exactly you build the windows installers or windows gui tools without using windows nor vs2008. If GCC/MINGW-w64 alone can deal with it, it would simplify the build process a lot (and helps to gain more contributors). > >I have tried MSYS2 but it refuse to build. Only in a VM with the relic VS2008 SP1 installed can generate the installer. > >If the build process is clear (and hopefully not requiring C#), than writing a cmake build script MIGHT be possible.
1.0
Building with the MinGW toolchain - https://github.com/SoftEtherVPN/SoftEtherVPN/issues/65#issuecomment-373994100 > I'd also like to know how exactly you build the windows installers or windows gui tools without using windows nor vs2008. If GCC/MINGW-w64 alone can deal with it, it would simplify the build process a lot (and helps to gain more contributors). > >I have tried MSYS2 but it refuse to build. Only in a VM with the relic VS2008 SP1 installed can generate the installer. > >If the build process is clear (and hopefully not requiring C#), than writing a cmake build script MIGHT be possible.
non_priority
building with the mingw toolchain i d also like to know how exactly you build the windows installers or windows gui tools without using windows nor if gcc mingw alone can deal with it it would simplify the build process a lot and helps to gain more contributors i have tried but it refuse to build only in a vm with the relic installed can generate the installer if the build process is clear and hopefully not requiring c than writing a cmake build script might be possible
0
264,599
23,126,989,002
IssuesEvent
2022-07-28 06:53:44
void-linux/void-packages
https://api.github.com/repos/void-linux/void-packages
opened
xfreerdp: SmartScaling requested but compiled without libcairo support
bug needs-testing
### Is this a new report? Yes ### System Info Void 5.18.11_1 x86_64 GenuineIntel notuptodate rFFFFFFFFFFFFFF ### Package(s) Affected freerdp-2.7.0_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? _No response_ ### Expected behaviour freerdp should support `/smart-sizing` to adjust resolution to window size. Otherwise, there is no use on 4k screens. ### Actual behaviour when used `freerdp /smart-sizing ...` command, I am getting the followoing error `SmartScaling requested but compiled without libcairo support` ### Steps to reproduce just run this command with RDP host `xfreerdp /u:Administrator /p:"password" /v:<host> /smart-sizing /scale:100 /w:1920 /h:1080`
1.0
xfreerdp: SmartScaling requested but compiled without libcairo support - ### Is this a new report? Yes ### System Info Void 5.18.11_1 x86_64 GenuineIntel notuptodate rFFFFFFFFFFFFFF ### Package(s) Affected freerdp-2.7.0_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? _No response_ ### Expected behaviour freerdp should support `/smart-sizing` to adjust resolution to window size. Otherwise, there is no use on 4k screens. ### Actual behaviour when used `freerdp /smart-sizing ...` command, I am getting the followoing error `SmartScaling requested but compiled without libcairo support` ### Steps to reproduce just run this command with RDP host `xfreerdp /u:Administrator /p:"password" /v:<host> /smart-sizing /scale:100 /w:1920 /h:1080`
non_priority
xfreerdp smartscaling requested but compiled without libcairo support is this a new report yes system info void genuineintel notuptodate rffffffffffffff package s affected freerdp does a report exist for this bug with the project s home upstream and or another distro no response expected behaviour freerdp should support smart sizing to adjust resolution to window size otherwise there is no use on screens actual behaviour when used freerdp smart sizing command i am getting the followoing error smartscaling requested but compiled without libcairo support steps to reproduce just run this command with rdp host xfreerdp u administrator p password v smart sizing scale w h
0
125,386
17,836,143,758
IssuesEvent
2021-09-03 01:32:47
kapseliboi/nuxt-client
https://api.github.com/repos/kapseliboi/nuxt-client
opened
CVE-2021-37713 (High) detected in tar-6.0.5.tgz, tar-4.4.13.tgz
security vulnerability
## CVE-2021-37713 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-6.0.5.tgz</b>, <b>tar-4.4.13.tgz</b></p></summary> <p> <details><summary><b>tar-6.0.5.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.0.5.tgz">https://registry.npmjs.org/tar/-/tar-6.0.5.tgz</a></p> <p>Path to dependency file: nuxt-client/package.json</p> <p>Path to vulnerable library: nuxt-client/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - node-sass-5.0.0.tgz (Root Library) - node-gyp-7.1.2.tgz - :x: **tar-6.0.5.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p> Dependency Hierarchy: - vuepress-1.8.2.tgz (Root Library) - core-1.8.2.tgz - chokidar-2.1.8.tgz - fsevents-1.2.12.tgz - node-pre-gyp-0.14.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37713 (High) detected in tar-6.0.5.tgz, tar-4.4.13.tgz - ## CVE-2021-37713 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-6.0.5.tgz</b>, <b>tar-4.4.13.tgz</b></p></summary> <p> <details><summary><b>tar-6.0.5.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.0.5.tgz">https://registry.npmjs.org/tar/-/tar-6.0.5.tgz</a></p> <p>Path to dependency file: nuxt-client/package.json</p> <p>Path to vulnerable library: nuxt-client/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - node-sass-5.0.0.tgz (Root Library) - node-gyp-7.1.2.tgz - :x: **tar-6.0.5.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p> Dependency Hierarchy: - vuepress-1.8.2.tgz (Root Library) - core-1.8.2.tgz - chokidar-2.1.8.tgz - fsevents-1.2.12.tgz - node-pre-gyp-0.14.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in tar tgz tar tgz cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar tgz tar for node library home page a href path to dependency file nuxt client package json path to vulnerable library nuxt client node modules tar package json dependency hierarchy node sass tgz root library node gyp tgz x tar tgz vulnerable library tar tgz tar for node library home page a href dependency hierarchy vuepress tgz root library core tgz chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch main vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted this is in part accomplished by sanitizing absolute paths of entries within the archive skipping archive entries that contain path portions and resolving the sanitized paths against the extraction target directory this logic was insufficient on windows systems when extracting tar files that contained a path that was not an absolute path but specified a drive letter different from the extraction target such as c some path if the drive letter does not match the extraction target for example d extraction dir then the result of path resolve extractiondirectory entrypath would resolve against the current working directory on the c drive rather than the extraction target directory additionally a portion of the path could occur immediately after the drive letter such as c foo and was not properly sanitized by the logic that checked for within the normalized and split portions of the path this only affects users of node tar on windows systems these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar there is no reasonable way to work around this issue without performing the same path normalization procedures that node tar now does users are encouraged to upgrade to the latest patched versions of node tar rather than attempt to sanitize paths themselves publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0