id
int64 393k
2.82B
| repo
stringclasses 68
values | title
stringlengths 1
936
| body
stringlengths 0
256k
โ | labels
stringlengths 2
508
| priority
stringclasses 3
values | severity
stringclasses 3
values |
---|---|---|---|---|---|---|
2,674,581,831 |
flutter
|
fllutter create should initialize analytics_options.yaml with stricter settings
|
### Use case
When creating a new flutter project with `fluter create` the generated `analysis_options.yaml` does only contain
`flutter_lints`
### Proposal
It should at least set
```yaml
analyzer:
language:
strict-casts: true
strict-inference: true
strict-raw-types: true
```
also some lints should be set to level `errror` see
https://github.com/dart-lang/sdk/issues/57098#issuecomment-2485661859
|
c: new feature,package,c: proposal,p: flutter_lints
|
low
|
Minor
|
2,674,610,621 |
storybook
|
Luminosity contrast ratio of keyboard focus indicator with the grey background on the left navigation panel is less than minimum required ratio of 3:1.
|
### Describe the bug
Luminosity contrast ratio of keyboard focus indicator with the grey background on the left navigation panel is less than minimum required ratio of 3:1
**Observation:**
Same issue is repro'ing on both Mac and Windows.
**User Impact:**
Visual impaired users and low vision users will be impacted if Luminosity contrast ratio of keyboard focus indicator is less than the required ratio due to which they will not be able to identify their focus properly and will not be able to complete their task easily.
Actual Result:
Luminosity contrast ratio of keyboard focus indicator with the grey background for 'Content' Control in the left navigation panel is 1.2:1 which is less than minimum required ratio of 3:1.
Refer Attachment: Luminosity contrast ratio of keyboard focus indicator with the grey background is less than minimum required.

Expected Result:
Luminosity contrast ratio of keyboard focus indicator with the grey background for all the controls in the left navigation panel should be greater than minimum required ratio of 3:1.
### Reproduction link
About / Introduction - Page โ
Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net)
### Reproduction steps
1. Open URL: [About / Introduction - Page โ
Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net)](https://ambitious-cliff-0c8148010.2.azurestaticapps.net/?path=/story/about-introduction--page) in latest edge browser.
2. Press tab key to reach to the controls in the left pane.
3. Press tab key and reach to the 'Content' control.
4. Open Color contrast analyzer and select foreground and background for the focus indicator using color picker.
5. Verify whether on pressing tab keys to reach to the 'Contents' control present in the left navigation panel, luminosity contrast ratio of keyboard focus indicator with the grey background is greater than minimum required or 3:1 or not.
### System
```bash
Test Environment:
OS: Windows 11 Version 24H2 OS build 26100.1742
Browser: Microsoft New Edge Version129.0.2792.52(Official build) (64-bit)
URL: About / Introduction - Page โ
Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net)
Tool Used: Color Contrast Analyzer
Mac Test Environment:
OS: macOS Sequoia 15.0.1
Browser: Microsoft New Edge Version130.0.2849.56 (Official build) (64-bit)
RL: About / Introduction - Page โ
Storybook (ambitious-cliff-0c8148010.2.azurestaticapps.net)
Tool Used: Color Contrast Analyzer
```
### Additional context
_No response_
|
bug,accessibility
|
low
|
Critical
|
2,674,611,675 |
pytorch
|
DISABLED test_sync (__main__.TestMPS)
|
Platforms: mac, macos
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_sync&suite=TestMPS&limit=100) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/33237372365).
Over the past 3 hours, it has been determined flaky in 4 workflow(s) with 10 failures and 4 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_sync`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_mps.py`
cc @clee2000 @malfet @albanD @kulinseth @DenisVieriu97 @jhavukainen
|
triaged,module: flaky-tests,module: macos,skipped,module: mps
|
low
|
Critical
|
2,674,675,963 |
tensorflow
|
Aborted (core dumped) in `tf.raw_ops.MatrixSolve`
|
### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.17
### Custom code
Yes
### OS platform and distribution
Linux Ubuntu 22.04.3 LTS (x86_64)
### Mobile device
_No response_
### Python version
3.9.13
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
When the shape of the matrix argument is empty and the gpu is available, tf.raw_ops.MatrixSolve triggers a crash.
It can be reproduced on tf-nightly when the gpu is available.
### Standalone code to reproduce the issue
```shell
import tensorflow as tf
tf.raw_ops.MatrixSolve(matrix=tf.random.uniform([], dtype=tf.dtypes.double, maxval=1000000000), rhs=tf.random.uniform([1, 2], dtype=tf.dtypes.double, maxval=1000000000), adjoint=True)
```
### Relevant log output
```shell
2024-11-20 14:51:08.714846: I tensorflow/core/util/port.cc:153] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.
2024-11-20 14:51:08.775383: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:485] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
2024-11-20 14:51:08.852267: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:8454] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
2024-11-20 14:51:08.876168: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1452] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
2024-11-20 14:51:08.931206: I tensorflow/core/platform/cpu_feature_guard.cc:210] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.
To enable the following instructions: SSE4.1 SSE4.2 AVX AVX2 AVX512F AVX512_VNNI AVX512_BF16 AVX512_FP16 AVX_VNNI AMX_TILE AMX_INT8 AMX_BF16 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.
2024-11-20 14:51:16.385650: I tensorflow/core/common_runtime/gpu/gpu_device.cc:2021] Created device /job:localhost/replica:0/task:0/device:GPU:0 with 535 MB memory: -> device: 0, name: NVIDIA GeForce RTX 4090, pci bus id: 0000:1f:00.0, compute capability: 8.9
2024-11-20 14:51:16.387914: I tensorflow/core/common_runtime/gpu/gpu_device.cc:2021] Created device /job:localhost/replica:0/task:0/device:GPU:1 with 71 MB memory: -> device: 1, name: NVIDIA GeForce RTX 4090, pci bus id: 0000:d4:00.0, compute capability: 8.9
2024-11-20 14:51:16.542383: F tensorflow/core/framework/tensor_shape.cc:356] Check failed: d >= 0 (0 vs. -1)
Aborted (core dumped)
```
|
stat:awaiting tensorflower,type:bug,comp:ops,2.17
|
medium
|
Critical
|
2,674,687,480 |
pytorch
|
[AOTInductor] AOT compiled models do not consistently check input shapes
|
### ๐ Describe the bug
AOT compiled models do not consistently provide the same runtime input Tensor checks and safety as do either plain PyTorch models or PyTorch exported graph models (via `torch.export`). Oddly, whether these checks happen with AOT models seems to depend on the shape of the data used to export the PyTorch model (???).
In the repro below, I export a toy model as a graph and then as an AOT compiled model and run inference on both. In the first case, I seed the "happy path" model with 2D input and both the torch.export'ed and AOT compiled models expectedly raise `RuntimeException`s and complain about input Tensor shape mismatch. However, in the second case, the "sad path", the model is exported with 1D example data and the AOT compiled model **does not** error on invalid input data. Inference completes successfully which is **not** the expected behavior -- I'd expected the AOT compiled model to raise a RuntimeException as it did in the "happy path".
**repro**
```python
import torch
class Model(torch.nn.Module):
def __init__(self) -> None:
super().__init__()
self._layer = torch.nn.Linear(2, 1)
def forward(self, X: torch.Tensor) -> torch.Tensor:
return self._layer(X)
model = Model()
# HAPPY PATH: 2D example data
print("-" * 50 + "\nHAPPY PATH")
with torch.inference_mode():
device = "cuda:0"
# device = "cpu"
X = torch.rand((1, 2), device=device)
model = model.to(device)
# AOTI compile on CPU
example_inputs = X
exported_program = torch.export.export(model, (example_inputs,),)
exported_model = exported_program.module().to(device)
so_path = torch._inductor.aot_compile(exported_model, (example_inputs,),)
aot_model = torch._export.aot_load(so_path, device)
# Inference: OK
print(">> OK: Exported Graph Model: ", exported_model(X))
print(">> OK: AOT Compiled Model: ", aot_model(X))
# Inference: FAILS
# NOTE: shape of X_t is invalid
X_t = torch.rand((3, 1), device=device)
try:
exported_model(X_t)
except RuntimeError as exc:
print(">> FAIL: Exported Graph Model: ", exc)
try:
aot_model(X_t)
except RuntimeError as exc:
print(">> FAIL: AOT Compiled Model: ", exc)
print("-" * 50)
print("-" * 50 + "\nSAD PATH")
# SAD PATH: 1D example data
with torch.inference_mode():
device = "cuda:0"
# make example data 1D
X = torch.rand((2,), device=device)
model = model.to(device)
# AOTI compile on CPU
example_inputs = X
exported_program = torch.export.export(model, (example_inputs,),)
exported_model = exported_program.module().to(device)
so_path = torch._inductor.aot_compile(exported_model, (example_inputs,),)
aot_model = torch._export.aot_load(so_path, device)
# Inference: OK
print(">> OK: Exported Graph Model: ", exported_model(X))
print(">> OK: AOT Compiled Model: ", aot_model(X))
# Inference: FAILS
# NOTE: shape of X_t is invalid
X_t = torch.rand((3, 1), device=device)
try:
exported_model(X_t)
except RuntimeError as exc:
print(">> FAIL: Exported Graph Model: ", exc)
print(">> THIS SHOULD FAIL BUT IS OK: AOT Compiled Model: ", aot_model(X_t))
print("-" * 50)
```
**output**
```bash
--------------------------------------------------
HAPPY PATH
>> OK: Exported Graph Model: (tensor([[-0.0415]], device='cuda:0'),)
>> OK: AOT Compiled Model: tensor([[-0.0415]], device='cuda:0')
>> FAIL: Exported Graph Model: Expected input at *args[0].shape[0] to be equal to 1, but got 3
Error: mat1 and mat2 shapes cannot be multiplied (3x1 and 2x1)
Exception raised from meta at ../aten/src/ATen/native/LinearAlgebra.cpp:189 (most recent call first):
...
>> FAIL: AOT Compiled Model: run_func_( container_handle_, input_handles.data(), input_handles.size(), output_handles.data(), output_handles.size(), cuda_stream_handle, proxy_executor_handle_) API call failed at ../torch/csrc/inductor/aoti_runner/model_container_runner.cpp, line 75
--------------------------------------------------
--------------------------------------------------
SAD PATH
>> OK: Exported Graph Model: (tensor([-0.0204], device='cuda:0'),)
>> OK: AOT Compiled Model: tensor([-0.0204], device='cuda:0')
>> FAIL: Exported Graph Model: Unexpected number of dimensions in input at *args[0].shape (expected torch.Size([2]), got torch.Size([3, 1]))
>> THIS SHOULD FAIL BUT IS OK: AOT Compiled Model: tensor([-0.0676], device='cuda:0')
--------------------------------------------------
```
This was observed on CUDA + `torch==2.4.1,2.5.1,nightly`.
### Versions
```bash
PyTorch version: 2.4.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 24.04 LTS (x86_64)
GCC version: (Ubuntu 13.2.0-23ubuntu4) 13.2.0
Clang version: Could not collect
CMake version: version 3.28.3
Libc version: glibc-2.39
Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0] (64-bit runtime)
Python platform: Linux-6.8.0-1018-aws-x86_64-with-glibc2.39
Is CUDA available: True
CUDA runtime version: 12.0.140
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA L40S
Nvidia driver version: 560.35.03
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 48 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 4
On-line CPU(s) list: 0-3
Vendor ID: AuthenticAMD
Model name: AMD EPYC 7R13 Processor
CPU family: 25
Model: 1
Thread(s) per core: 2
Core(s) per socket: 2
Socket(s): 1
Stepping: 1
BogoMIPS: 5300.00
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf tsc_known_freq pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch topoext ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 invpcid rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr rdpru wbnoinvd arat npt nrip_save vaes vpclmulqdq rdpid
Hypervisor vendor: KVM
Virtualization type: full
L1d cache: 64 KiB (2 instances)
L1i cache: 64 KiB (2 instances)
L2 cache: 1 MiB (2 instances)
L3 cache: 8 MiB (1 instance)
NUMA node(s): 1
NUMA node0 CPU(s): 0-3
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; IBRS_FW; STIBP always-on; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] numpy==2.1.2
[pip3] nvidia-cublas-cu12==12.1.3.1
[pip3] nvidia-cuda-cupti-cu12==12.1.105
[pip3] nvidia-cuda-nvrtc-cu12==12.1.105
[pip3] nvidia-cuda-runtime-cu12==12.1.105
[pip3] nvidia-cudnn-cu12==9.1.0.70
[pip3] nvidia-cufft-cu12==11.0.2.54
[pip3] nvidia-curand-cu12==10.3.2.106
[pip3] nvidia-cusolver-cu12==11.4.5.107
[pip3] nvidia-cusparse-cu12==12.1.0.106
[pip3] nvidia-nccl-cu12==2.20.5
[pip3] nvidia-nvjitlink-cu12==12.6.68
[pip3] nvidia-nvtx-cu12==12.1.105
[pip3] onnxruntime-gpu==1.20.0
[pip3] torch==2.4.1
[pip3] triton==3.0.0
[conda] Could not collect
```
cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4
|
oncall: pt2,oncall: export
|
low
|
Critical
|
2,674,711,373 |
godot
|
handle on exported range property detaches from slider when scrolling while control is hovered
|
### Tested versions
Found in v4.3.stable.official [77dcf97d8]
### System information
Godot v4.3.stable - Ubuntu 22.04.5 LTS 22.04 - X11 - Vulkan (Forward+) - dedicated NVIDIA GeForce GTX 970 (nvidia; 550.120) - Intel(R) Core(TM) i9-10900K CPU @ 3.70GHz (20 Threads)
### Issue description
This is a visual bug. the slider handles detach from the sliders when hovered while scrolling with the scrollbar
https://github.com/user-attachments/assets/b7efe512-a994-4c14-8b30-0212a9dd4737
### Steps to reproduce
1. go to the inspector on any node or resource with an exported float range property
2. make scrollbar appear and hold click on the scrollbar
3. move mouse over the range slider while scrolling up/down
4. notice the handle detaches
### Minimal reproduction project (MRP)
Empty project
|
bug,topic:gui
|
low
|
Critical
|
2,674,714,085 |
ollama
|
OLLAMA_MAX_QUEUE does not limit requests to the same model
|
### What is the issue?
It seems that OLLAMA_MAX_QUEUE is not taking effect. My environment is Windows 11, and I have set OLLAMA_NUM_PARALLEL=1,
set OLLAMA_MAX_QUEUE=1, but excessive requests are still queuing up instead of returning an error.
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.14
|
feature request
|
low
|
Critical
|
2,674,734,144 |
PowerToys
|
Copy from PowerToys Run error
|
### Microsoft PowerToys version
0.86.0
### Installation method
WinGet
### Running as admin
No
### Area(s) with issue?
PowerToys Run
### Steps to reproduce
Open PowerToys Run, insert some equation. Click on the "Copy this number to clipboard". Error pops up with following:

More info in the following bug report summary:
[PowerToysReport_2024-11-20-07-38-11.zip](https://github.com/user-attachments/files/17826692/PowerToysReport_2024-11-20-07-38-11.zip)
### โ๏ธ Expected Behavior
Copy value to clipboard
### โ Actual Behavior
_No response_
### Other Software
_No response_
|
Issue-Bug,Needs-Triage
|
low
|
Critical
|
2,674,833,953 |
pytorch
|
How to fuse batchnorm to conv2d in the graph exported by torch.export
|
I used the torch.export to export my CNN model in eval mode,but the op batchnorm still exists. how to eliminate it. Is there some options in torch.export.export function or I should write a fusion pass by myself.
Thanks.
code:
```
import torch
import torch.nn as nn
class CNN(nn.Module):
def __init__(self):
super().__init__()
self.conv1 = nn.Conv2d(in_channels=16, out_channels=16, kernel_size=3, stride=1, padding=1)
self.bn1 = nn.BatchNorm2d(16)
def forward(self, x):
x = self.conv1(x)
x = self.bn1(x)
return x
torch.manual_seed(0)
model=CNN().eval()
input=torch.randn(3,16,224,224)
ep=torch.export.export(model,(input,))
print(ep.graph)
```
graph:
```
graph():
%p_conv1_weight : [num_users=1] = placeholder[target=p_conv1_weight]
%p_conv1_bias : [num_users=1] = placeholder[target=p_conv1_bias]
%p_bn1_weight : [num_users=1] = placeholder[target=p_bn1_weight]
%p_bn1_bias : [num_users=1] = placeholder[target=p_bn1_bias]
%b_bn1_running_mean : [num_users=1] = placeholder[target=b_bn1_running_mean]
%b_bn1_running_var : [num_users=1] = placeholder[target=b_bn1_running_var]
%b_bn1_num_batches_tracked : [num_users=0] = placeholder[target=b_bn1_num_batches_tracked]
%x : [num_users=1] = placeholder[target=x]
%conv2d : [num_users=1] = call_function[target=torch.ops.aten.conv2d.default](args = (%x, %p_conv1_weight, %p_conv1_bias, [1, 1], [1, 1]), kwargs = {})
%_native_batch_norm_legit_no_training : [num_users=1] = call_function[target=torch.ops.aten._native_batch_norm_legit_no_training.default](args = (%conv2d, %p_bn1_weight, %p_bn1_bias, %b_bn1_running_mean, %b_bn1_running_var, 0.1, 1e-05), kwargs = {})
%getitem : [num_users=1] = call_function[target=operator.getitem](args = (%_native_batch_norm_legit_no_training, 0), kwargs = {})
return (getitem,)
```
cc @chauhang @penguinwu @avikchaudhuri @gmagogsfm @zhxchen17 @tugsbayasgalan @angelayi @suo @ydwu4
|
oncall: pt2,oncall: export
|
low
|
Minor
|
2,674,880,248 |
vscode
|
Settings Editor Synchronization Error
|
<!-- โ ๏ธโ ๏ธ Do Not Delete This! bug_report_template โ ๏ธโ ๏ธ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- ๐ฎ Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- ๐ Search existing issues to avoid creating duplicates. -->
<!-- ๐งช Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- ๐ก Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- ๐ง Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- ๐ช If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- ๐ฃ Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.95.3
- OS Version: win 11
Steps to Reproduce:
1. Click the Settings button at the bottom left corner of the Workbench, which opens the Settings Editor window.
2. In the Settings Editor window, modify the visibility of the Workbench's Status Bar. (need)
3. Change the visibility of the status bar through the menu: View > Appearance > Status Bar.
4. At this point, the value of the status bar visibility in the Settings Editor window does not synchronize with the recently set state.
|
bug,settings-editor,confirmation-pending
|
low
|
Critical
|
2,674,901,219 |
rust
|
Compiled wasm32-wasip2 component from simple code requires excessive WASI interfaces
|
<!--
Thank you for filing a bug report! ๐ Please provide a short summary of the bug,
along with any information you feel relevant to replicating the bug.
-->
During my learning and experiments of wasip2, I tried this simple code and compiled it to a wasip2 component:
```rust
wit_bindgen::generate!({
// the name of the world in the `*.wit` input file
world: "formatter",
});
struct Formatter;
impl Guest for Formatter {
fn format_str(a: String, b: String) -> String {
let s = format!("{} + {}", a, b);
print(s.as_str());
s
}
}
export!(Formatter);
```
and
```
// format.wit
package component:formatter;
world formatter {
import print: func(s: string);
export format-str: func(a: string, b: string) -> string;
}
```
The host code, which runs the component, is:
```rust
use wasmtime::component::*;
use wasmtime::{Engine, Store};
use wasmtime_wasi::{WasiCtx, WasiImpl, WasiView};
use anyhow::Result;
// reference: https://docs.rs/wasmtime/latest/wasmtime/component/bindgen_examples/_0_hello_world/index.html
// reference: https://docs.wasmtime.dev/examples-rust-wasi.html
bindgen!({
path: "../implementation/wit/format.wit",
world: "formatter",
});
struct MyState {
// These two are required basically as a standard way to enable the impl of WasiView
wasi_ctx: WasiCtx,
table: ResourceTable,
}
impl WasiView for MyState {
fn table(&mut self) -> &mut ResourceTable {
&mut self.table
}
fn ctx(&mut self) -> &mut WasiCtx {
&mut self.wasi_ctx
}
}
impl FormatterImports for MyState {
fn print(&mut self, s: String) {
println!("{}", s);
}
}
/// copied from wasmtime_wasi::type_annotate, which is a private function
fn type_annotate<T: WasiView, F>(val: F) -> F
where
F: Fn(&mut T) -> WasiImpl<&mut T>,
{
val
}
fn main() -> Result<()> {
let engine = Engine::default();
let component = Component::from_file(
&engine,
"../implementation/target/wasm32-wasip2/release/implementation.wasm",
)?;
let mut linker = Linker::new(&engine);
let ctx = wasmtime_wasi::WasiCtxBuilder::new().build();
let state = MyState {
wasi_ctx: ctx,
table: ResourceTable::new(),
};
let mut store = Store::new(&engine, state);
Formatter::add_to_linker(&mut linker, |s| s)?;
// Note:
// The below block is copied from `wasmtime_wasi::add_to_linker_sync`.
// why a "format!()" in the implementation needs `sync::filesystem::types`, `sync::io::streams`, `cli::exit`, `cli::environment`, `cli::stdin`, `cli::stdout`, `cli::stderr`?
{
let l = &mut linker;
let closure = type_annotate::<MyState, _>(|t| WasiImpl(t));
wasmtime_wasi::bindings::sync::filesystem::types::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::filesystem::preopens::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::io::error::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::sync::io::streams::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::cli::exit::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::cli::environment::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::cli::stdin::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::cli::stdout::add_to_linker_get_host(l, closure)?;
wasmtime_wasi::bindings::cli::stderr::add_to_linker_get_host(l, closure)?;
}
let bindings = Formatter::instantiate(&mut store, &component, &linker)?;
let result = bindings.call_format_str(&mut store, "a", "b")?;
println!("format_str: {}", result);
Ok(())
}
```
The block with a note binds many interfaces to avoid runtime errors that says something like
`component imports instance wasi:cli/[email protected], but a matching implementation was not found in the linker.` Removing any one of the lines in the block will result in a runtime error.
I expect this compiled component requires none of these WASI interfaces to run, since it has nothing to do with io, cli, etc. Binding these unnecessary interfaces may raise security concerns.
The full minimized code is [here](https://github.com/ifsheldon/wit_interface_issue).
As a kind person pointed out on [ByteAlliance Zulip](https://bytecodealliance.zulipchat.com/#narrow/channel/217126-wasmtime/topic/Help.3A.20unnecessary.20interface.20requirement.20for.20using.20format!.28.29.3F/near/483417497), these interfaces are required by `std`.
Probably there's a way to minimize or prune the interface requirements in the compilation? I think rustc has all the information of which effects are used by any one of functions/macros that is used by a user.
At the very lease, I think we should document these requirements somewhere, so there are no hidden/dark interface dependencies that are not specified and unknown in WIT files.
### Meta
<!--
If you're using the stable version of the compiler, you should also check if the
bug also exists in the beta or nightly versions.
-->
`rustc --version --verbose`:
```
rustc 1.82.0 (f6e511eec 2024-10-15)
binary: rustc
commit-hash: f6e511eec7342f59a25f7c0534f1dbea00d01b14
commit-date: 2024-10-15
host: aarch64-apple-darwin
release: 1.82.0
LLVM version: 19.1.1
```
|
T-libs,C-discussion,O-wasi
|
medium
|
Critical
|
2,674,935,089 |
tensorflow
|
Does tfl.quantize support QI4 data type
|
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 22.04
- TensorFlow installed from (source or binary): source
- TensorFlow version (or github SHA if from source): the latest version
**Standalone code to reproduce the issue**
The [doc of tfl.quantize](https://www.tensorflow.org/mlir/tfl_ops#tflquantize_tflquantizeop) supports QI4 data types, but [the kernel implementation](https://source.chromium.org/chromium/chromium/src/+/main:third_party/tflite/src/tensorflow/lite/kernels/quantize.cc;l=121) doesn't support QI4 data type, the implementation seems not be consistent with the doc.
|
stat:awaiting tensorflower,type:support,comp:lite
|
medium
|
Minor
|
2,675,055,650 |
flutter
|
Linux plugin_test is 2.11% flaky
|
<!-- meta-tags: To be used by the automation script only, DO NOT MODIFY.
{
"name": "Linux plugin_test"
}
-->
The post-submit test builder `Linux plugin_test` had a flaky ratio 2.11% for the past (up to) 100 commits, which is above our 2.00% threshold.
One recent flaky example for a same commit: https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21536
Commit: https://github.com/flutter/flutter/commit/638331b13328245d7fd5b8cffd729622e3518b8b
Flaky builds:
https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21536
https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21529
https://ci.chromium.org/ui/p/flutter/builders/prod/Linux%20plugin_test/21528
Recent test runs:
https://flutter-dashboard.appspot.com/#/build?taskFilter=Linux%20plugin_test
Please follow https://github.com/flutter/flutter/blob/master/docs/infra/Reducing-Test-Flakiness.md#fixing-flaky-tests to fix the flakiness and enable the test back after validating the fix (internal dashboard to validate: go/flutter_test_flakiness).
|
platform-android,P1,c: flake,team-android,triaged-android
|
medium
|
Major
|
2,675,235,883 |
terminal
|
Releasing both pressed Shift keys (left+right) reports only one
|
### Windows Terminal version
current main
### Windows build number
10.0.19045.5011
### Other Software
_No response_
### Steps to reproduce
1. Activate "debug tap" pane.
- Add the global setting (near the top of your JSON file):
```
"debugFeatures": true
```
- Open a new tab while holding down both left and right Alt .
2. Press and release both Shift keys simultaneously.
### Expected Behavior
There are four reports in the "debug tap" pane:
- The first Shift key was pressed.
- The second Shift key was pressed.
- Some Shift key was released.
- The remaining Shift key was released.
### Actual Behavior

There are only three reports in the "debug tap" pane:
- The first Shift key was pressed.
- The second Shift key was pressed.
- Some Shift key was released.
|
Help Wanted,Area-Input,Issue-Bug,Product-Terminal,Priority-3
|
low
|
Critical
|
2,675,247,756 |
godot
|
Can't disable SSL in HttpRequest
|
### Tested versions
v4.4.dev4.mono.official [36e6207bb]
### System information
Godot v4.4.dev4.mono - Windows 10.0.26100 - Multi-window, 2 monitors - Vulkan (Forward+) - dedicated NVIDIA GeForce RTX 2070 (NVIDIA; 32.0.15.6094) - 13th Gen Intel(R) Core(TM) i7-13700 (24 threads)
### Issue description
I need post a https. for some reason, it will report TLS_HANDSHAKE_ERROR. i'm test in Godot and Qt. both will report this error.
then for Qt, i can add some code for disable tls:
```
QSslConfiguration sslConfig;
sslConfig.setPeerVerifyMode(QSslSocket::VerifyNone);
objRequest.setSslConfiguration(sslConfig);
```
then Qt can get response successful.
but for Godot, i'm try add code like this:
```
http_request.set_tls_options(TLSOptions.client_unsafe())
```
but unfortunately, it can't work. it always response TLS_HANDSHAKE_ERROR.
### Steps to reproduce
none
### Minimal reproduction project (MRP)
none
ADDTION: i'm search for this issue. but all of them was closed. maybe I'm use this in a wrong way, but i don't know how to do it.
|
bug,topic:network
|
low
|
Critical
|
2,675,312,557 |
vscode
|
Bug Report Must Be Improved
|
- There are lots of inappropriate issues
- Most of them are created via VSCode Issue Reporter screen
- The minimum description length is so small
- There is no minimum length requirement for Title, 1 character is enough
- For example, this is accepted: 
- Also the issue template can be improved in Github
- This is causing many issues be in bad form. For example: https://github.com/microsoft/vscode/issues/234040 https://github.com/microsoft/vscode/issues/234224 https://github.com/microsoft/vscode/issues/234196
|
feature-request,issue-reporter
|
low
|
Critical
|
2,675,349,542 |
flutter
|
ScrollAction slows down significantly when holding down keys
|
### Steps to reproduce
1. Place a [`MaterialApp`](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget at the root of your widget tree.
2. Place any scrollable widget below the [`MaterialApp`](https://api.flutter.dev/flutter/material/MaterialApp-class.html) and make sure it uses the [`PrimaryScrollController`](https://api.flutter.dev/flutter/widgets/PrimaryScrollController-class.html).
3. Try scrolling down by pressing ctrl+down for desktop or just the down arrow for web.
4. Then try doing the same, but holding the keys down.
### Expected results
I would expect to scroll faster when holding down the keys than when pressing them repeatedly.
### Actual results
Scrolling becomes significantly slower when holding down the keys compared to pressing them repeatedly.
### Code sample
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
void main() => runApp(const ExampleApp());
class ExampleApp extends StatelessWidget {
const ExampleApp({super.key});
@override
Widget build(BuildContext context) => MaterialApp(
home: Scaffold(
body: ListView.builder(
primary: true,
itemCount: 15,
itemBuilder: (BuildContext context, int index) => Container(
color: Colors.primaries[index],
height: 150,
child: Center(
child: Text(
'Entry $index',
style: Theme.of(context).textTheme.headlineMedium,
),
),
),
),
),
);
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
The first part shows the behavior when pressing the scroll keys repeatedly, and then while holding the scroll keys down.

</details>
### Logs
_No response_
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[โ] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale
en_US.UTF-8)
โข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter
โข Upstream repository https://github.com/flutter/flutter.git
โข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800
โข Engine revision a18df97ca5
โข Dart version 3.5.4
โข DevTools version 2.37.3
[โ] Android toolchain - develop for Android devices
โ Unable to locate Android SDK.
Install Android Studio from: https://developer.android.com/studio/index.html
On first launch it will assist you in installing the Android SDK components.
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
If the Android SDK has been installed to a custom location, please use
`flutter config --android-sdk` to update to that location.
[โ] Chrome - develop for the web
โข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable
[โ] Linux toolchain - develop for Linux desktop
โข clang version 19.1.0 (Fedora 19.1.0-1.fc41)
โข cmake version 3.30.5
โข ninja version 1.12.1
โข pkg-config version 2.3.0
[!] Android Studio (not installed)
โข Android Studio not found; download from https://developer.android.com/studio/index.html
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
[โ] VS Code (version 1.95.3)
โข VS Code at /usr/share/code
โข Flutter extension version 3.100.0
[โ] Connected device (2 available)
โข Linux (desktop) โข linux โข linux-x64 โข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64
โข Chrome (web) โข chrome โข web-javascript โข Brave Browser 131.1.73.89
[โ] Network resources
โข All expected network resources are available.
! Doctor found issues in 2 categories.
```
</details>
|
framework,f: scrolling,platform-web,a: desktop,P2,team-framework,triaged-framework
|
low
|
Critical
|
2,675,480,242 |
PowerToys
|
PowerToys RUN crash when I play game
|
### Microsoft PowerToys version
0.86.0
### Installation method
GitHub
### Running as admin
No
### Area(s) with issue?
PowerToys Run
### Steps to reproduce
I am playing DIRT RALLY 2 with boadless window.
Microsoft Windows [10.0.27749.1000]
### โ๏ธ Expected Behavior
I am starting race.
### โ Actual Behavior
game screen frozen and PowerToys RUN show error.
[2024-11-20.txt](https://github.com/user-attachments/files/17829455/2024-11-20.txt)
### Other Software
_No response_
|
Issue-Bug,Needs-Triage
|
low
|
Critical
|
2,675,501,977 |
tauri
|
[feat] How to modify the referer attribute of video and audio on the Android platform
|
### Describe the problem
How to modify the referer attribute of video and audio on the Android platform, because I am now making the app version of the website, but in Tauri, the referer looks like tarui.local
### Describe the solution you'd like
How to modify the referer attribute of video and audio on the Android platform, because I am now making the app version of the website, but in Tauri, the referer looks like tarui.local
### Alternatives considered
_No response_
### Additional context
_No response_
|
type: feature request
|
low
|
Minor
|
2,675,502,131 |
flutter
|
NestedScrollView's body is not scrollable using keyboard shortcuts with the ScrollAction
|
### Steps to reproduce
1. Place a [MaterialApp](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget at the root of your widget tree.
2. Place a [NestedScrollView](https://api.flutter.dev/flutter/widgets/NestedScrollView-class.html) below the [MaterialApp](https://api.flutter.dev/flutter/material/MaterialApp-class.html) widget.
3. Try scrolling down by pressing ctrl+down for desktop or just the down arrow for web.
### Expected results
I would expect the header and then the body of the nested scroll view to be scrolled down.
### Actual results
The header is scrolled down as expected, but when reaching its extent, the body is not scrolled.
### Code sample
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
void main() => runApp(const ExampleApp());
class ExampleApp extends StatelessWidget {
const ExampleApp({super.key});
@override
Widget build(BuildContext context) => MaterialApp(
home: Scaffold(
body: NestedScrollView(
headerSliverBuilder: (
BuildContext context,
bool innerBoxIsScrolled,
) =>
<Widget>[
SliverOverlapAbsorber(
handle: NestedScrollView.sliverOverlapAbsorberHandleFor(
context,
),
sliver: SliverAppBar(
pinned: true,
expandedHeight: 150.0,
forceElevated: innerBoxIsScrolled,
flexibleSpace: FlexibleSpaceBar(
title: const Text('SliverAppBar'),
background: Container(color: Colors.red),
centerTitle: true,
),
),
),
],
body: CustomScrollView(
slivers: <Widget>[
Builder(
builder: (context) => SliverOverlapInjector(
handle: NestedScrollView.sliverOverlapAbsorberHandleFor(
context,
),
),
),
SliverFixedExtentList(
itemExtent: 48.0,
delegate: SliverChildBuilderDelegate(
(BuildContext context, int index) => ListTile(
title: Text('Item $index'),
),
childCount: 30,
),
),
],
),
),
),
);
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
The following video shows the scrolling behavior when using a trackpad.

We'd expect similar scrolling behavior when using the arrow keys, but instead the scrolling stops when the scroll extent of the header is reached.

</details>
### Logs
_No response_
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[โ] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale
en_US.UTF-8)
โข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter
โข Upstream repository https://github.com/flutter/flutter.git
โข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800
โข Engine revision a18df97ca5
โข Dart version 3.5.4
โข DevTools version 2.37.3
[โ] Android toolchain - develop for Android devices
โ Unable to locate Android SDK.
Install Android Studio from: https://developer.android.com/studio/index.html
On first launch it will assist you in installing the Android SDK components.
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
If the Android SDK has been installed to a custom location, please use
`flutter config --android-sdk` to update to that location.
[โ] Chrome - develop for the web
โข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable
[โ] Linux toolchain - develop for Linux desktop
โข clang version 19.1.0 (Fedora 19.1.0-1.fc41)
โข cmake version 3.30.5
โข ninja version 1.12.1
โข pkg-config version 2.3.0
[!] Android Studio (not installed)
โข Android Studio not found; download from https://developer.android.com/studio/index.html
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
[โ] VS Code (version 1.95.3)
โข VS Code at /usr/share/code
โข Flutter extension version 3.100.0
[โ] Connected device (2 available)
โข Linux (desktop) โข linux โข linux-x64 โข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64
โข Chrome (web) โข chrome โข web-javascript โข Brave Browser 131.1.73.89
[โ] Network resources
โข All expected network resources are available.
! Doctor found issues in 2 categories.
```
</details>
|
framework,f: scrolling,platform-web,a: desktop,f: focus,has reproducible steps,P3,team-framework,triaged-framework,found in release: 3.24,found in release: 3.27
|
low
|
Minor
|
2,675,504,350 |
ui
|
[bug]: Can't install shadcn
|
### Describe the bug
npx shadcn@latest init -d
>> C:\Users\Hs\Documents\sam>
โ Preflight checks.
โ Verifying framework. Found Next.js.
โ Validating Tailwind CSS.
โ Validating import alias.
โ Writing components.json.
โ Checking registry.
โ Updating tailwind.config.js
โ Updating src\app\globals.css
โ Installing dependencies.
Something went wrong. Please check the error below for more details.
If the problem persists, please open an issue on GitHub.
Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge
spawn UNKNOWN
### Affected component/components
the component itself
### How to reproduce
npx shadcn@latest init -d
>> C:\Users\Hs\Documents\sam>
โ Preflight checks.
โ Verifying framework. Found Next.js.
โ Validating Tailwind CSS.
โ Validating import alias.
โ Writing components.json.
โ Checking registry.
โ Updating tailwind.config.js
โ Updating src\app\globals.css
โ Installing dependencies.
Something went wrong. Please check the error below for more details.
If the problem persists, please open an issue on GitHub.
Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge
spawn UNKNOWN
### Codesandbox/StackBlitz link
_No response_
### Logs
```bash
npx shadcn@latest init -d
>> C:\Users\Hs\Documents\sam>
โ Preflight checks.
โ Verifying framework. Found Next.js.
โ Validating Tailwind CSS.
โ Validating import alias.
โ Writing components.json.
โ Checking registry.
โ Updating tailwind.config.js
โ Updating src\app\globals.css
โ Installing dependencies.
Something went wrong. Please check the error below for more details.
If the problem persists, please open an issue on GitHub.
Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge
spawn UNKNOWN
```
### System Info
```bash
npx shadcn@latest init -d
>> C:\Users\Hs\Documents\sam>
โ Preflight checks.
โ Verifying framework. Found Next.js.
โ Validating Tailwind CSS.
โ Validating import alias.
โ Writing components.json.
โ Checking registry.
โ Updating tailwind.config.js
โ Updating src\app\globals.css
โ Installing dependencies.
Something went wrong. Please check the error below for more details.
If the problem persists, please open an issue on GitHub.
Command failed with UNKNOWN: yarn add tailwindcss-animate class-variance-authority lucide-react clsx tailwind-merge
spawn UNKNOWN
```
### Before submitting
- [X] I've made research efforts and searched the documentation
- [X] I've searched for existing issues
|
bug
|
low
|
Critical
|
2,675,508,933 |
deno
|
Deno.serve timeouts on AWS lambda
|
Version: Deno 2.0.6
I'm trying to deploy a **complied** deno project into AWS Lambda. I'm using Hono, and in the docs, they are using [Deno.serve](https://hono.dev/docs/getting-started/deno#_3-hello-world).
The problem is that lambda timeouts and returns `Internal Server Error`.
After investigating the problem it seems that the `Deno.serve` is causing this.
In the log I can see:
```
2024-11-20T09:22:13.960Z | Listening on http://0.0.0.0:8000/
2024-11-20T09:22:22.838Z | EXTENSION Name: lambda-adapter State: Ready Events: []
2024-11-20T09:22:22.838Z | INIT_REPORT Init Duration: 9999.92 ms Phase: init Status: timeout
```
Steps:
1. Create a `main.ts` with `Deno.serve((req) => new Response("Hello World!"));`
2. Build & deploy the app:
```
deno compile -A --target x86_64-unknown-linux-gnu --no-check -o bootstrap main.ts && chmod +x ./bootstrap && zip -r -D artifact.zip bootstrap && aws lambda update-function-code --region eu-central-1 --function-name deno-compiled --zip-file fileb://artifact.zip
```
3. Remember to add lambda layer adapter `arn:aws:lambda:eu-central-1:753240598075:layer:LambdaAdapterLayerX86:23`.
I know there is a dockerised approach for deploying Deno, but I don't want to use it because it uses too much memory.
If you take an oak example from [aws-lambda-web-adapter examples](https://github.com/awslabs/aws-lambda-web-adapter/blob/main/examples/deno-zip/src/main.ts), compile it, and deploy it to AWS Lambda - it works.
|
needs investigation
|
low
|
Critical
|
2,675,606,238 |
TypeScript
|
NoInfer not working on callback parameter destructuring
|
### ๐ Search Terms
NoInfer parameter callback destructuring
### ๐ Version & Regression Information
Latest to date.
### โฏ Playground Link
https://www.typescriptlang.org/play/?#code/MYGwhgzhAEBiD28A8AVANNASgPmgbwCgBIYeAOwgBcAnAV2EvmoApqBTMAE3JAE9pgAIwBc0ZpVHpoAfVEA5eAEkyAMzbUkeaAA9RZWgFtB6gNzR4lABbr5S1eqQ5oAX2wBKaAF5cmD3mfEAPSBAuRUdAxMrBzcZHwCImIS0FKy+Dp6hsbUZhbW1LbKahpOzh7eWH4BRMGhFDT0jCzsXDz8QqLikhhpCkUOWrrQ+kam5lY2WC7uXj5VBAEEpPUyAIxew2wA7nCIzMxkYAZsouEAlmQA5hiDLuW4zFqHxxjad25mtQjI51cYtGQANZkeBbMjYAi1CCWeC0ECcaDGXY-GgXa7pZ4naC-S5mIYjbLTAhLMKUGQAJg2ZG2yP2mNOqKu9zETyObHenxCWyYgIgom+SBxN2GbIZ1DRRKAA
### ๐ป Code
```ts
class Foo<T, R> {
constructor(readonly cb: (t: T, _: NoInfer<{ x: number; other: NoInfer<R> }>) => R) {}
}
const _1 = new Foo((name: string, { x }) => ({ name, x })); // Foo<string, unknown>
// should be Foo<string, { name: string; x: number }>
const _2 = new Foo((name: string) => ({ name })); // works: Foo<string, { name: string }>
```
### ๐ Actual behavior
infers `unknown`.
### ๐ Expected behavior
Should ignore the parameter tagged with `NoInfer` and infer correctly.
### Additional information about the issue
Is this intended behavior? feels like a bug to me.
If the callback parameter is destructured partially it infers `unknown`, but it works if the parameter is omitted.
I tried `NoInfer` at various locations.
|
Help Wanted,Possible Improvement
|
medium
|
Critical
|
2,675,669,443 |
pytorch
|
[Dynamo] Dynamo is not aware of the function users override
|
### ๐ Describe the bug
Hi, we are trying to enabling Llama in Intel HPU, but found graph break in `torch.cuda.current_device()` . We have a package GPU migration tool, it replaces `torch.cuda.current_device()` to `torch.hpu.current_device()`. But we found dynamo is not aware of the function override, and still consider `torch.cuda.current_device()`, eventually leading to graph break or program failed.
Can you give us some suggestions or have any plan to fix it?
### Error logs
Here is a simple case
```python
import os
os.environ["PT_HPU_LAZY_MODE"] = "0"
import torch
import habana_frameworks.torch.dynamo.compile_backend
def hpu_device():
return torch.hpu.current_device()
torch.cuda.current_device = hpu_device
@torch.compile(backend="hpu_backend", fullgraph=True)
def foo(x):
dim_size = list(x.size())
y = torch.tensor([6], device=torch.cuda.current_device())
return torch.mul(x, y)
x = torch.tensor([6], device=torch.hpu.current_device())
foo(x)
```
Our GPU migration tool implement, override cuda current_device():
```python
@CudaModuleRegister.register_f()
def current_device() -> int:
"""
.. py:gpumgrcall:: current_device.hpu_match
Maps torch.cuda.current_device to torch.hpu.current_device.
"""
G_LOGGER.info(
api_type="hpu_match",
func_prefix="torch.cuda",
new_call="torch.hpu.current_device()",
)
return torch.hpu.current_device()
```
The error gets from dynamo:
```
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/bytecode_transformation.py", line 1322, in transform_code_object
transformations(instructions, code_options)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 219, in _fn
return fn(*args, **kwargs)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/convert_frame.py", line 634, in transform
tracer.run()
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 2796, in run
super().run()
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 983, in run
while self.step():
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 895, in step
self.dispatch_table[inst.opcode](self, inst)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 582, in wrapper
return inner_fn(self, inst)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 1602, in CALL_FUNCTION
self.call_function(fn, args, {})
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/symbolic_convert.py", line 830, in call_function
self.push(fn.call_function(self, args, kwargs)) # type: ignore[arg-type]
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/variables/functions.py", line 727, in call_function
unimplemented(msg)
File "/usr/local/lib/python3.10/dist-packages/torch/_dynamo/exc.py", line 297, in unimplemented
raise Unsupported(msg, case_name=case_name)
torch._dynamo.exc.Unsupported: 'skip function hpu_device in file /root/test.py'
from user code:
File "/root/test.py", line 14, in foo
y = torch.tensor([6], device=torch.cuda.current_device())
```
### Versions
```
PyTorch version: 2.5.1a0+git3ba5690
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.5 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: version 3.22.1
Libc version: glibc-2.35
Python version: 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0] (64-bit runtime)
Python platform: Linux-5.15.0-122-generic-x86_64-with-glibc2.35
Is CUDA available: False
CUDA runtime version: No CUDA
CUDA_MODULE_LOADING set to: N/A
GPU models and configuration: No CUDA
Nvidia driver version: No CUDA
cuDNN version: No CUDA
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 46 bits physical, 57 bits virtual
Byte Order: Little Endian
CPU(s): 160
On-line CPU(s) list: 0-159
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) Platinum 8380 CPU @ 2.30GHz
CPU family: 6
Model: 106
Thread(s) per core: 2
Core(s) per socket: 40
Socket(s): 2
Versions of relevant libraries:
[pip3] habana-torch-dataloader==1.19.0.428
[pip3] habana-torch-plugin==1.19.0.428
[pip3] numpy==1.23.5
[pip3] pytorch-lightning==2.4.0
[pip3] torch==2.5.1a0+git3ba5690
[pip3] torch_tb_profiler==0.4.0
[pip3] torchaudio==2.5.1a0+1661daf
[pip3] torchdata==0.9.0+d4bb3e6
[pip3] torchmetrics==1.6.0
[pip3] torchtext==0.18.0a0+9bed85d
[pip3] torchvision==0.20.1a0+3ac97aa
[conda] Could not collect
```
cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @amjames
|
triaged,oncall: pt2,module: dynamo
|
low
|
Critical
|
2,675,693,600 |
flutter
|
Unable to scroll diagonally with a trackpad in a TwoDimensionalScrollView
|
### Steps to reproduce
1. Build the example application from the ["2D scrolling - Flutter Build Show"](https://www.youtube.com/watch?v=ppEdTo-VGcg).
2. Try scrolling diagonally using a trackpad.
### Expected results
I would expect the viewport to change offset in both axes simultaneously.
### Actual results
The viewport changes its offset in a single axis only.
### Code sample
<details open><summary>Code sample</summary>
[DartPad - Two Dimensional Grid in Flutter](https://dartpad.dev/?id=4424936c57ed13093eb389123383e894)
You can additionaly wrap the scroll view in a [Listener](https://api.flutter.dev/flutter/widgets/Listener-class.html) to make sure the scroll events have offset in both axes.
```dart
...
body: Listener(
onPointerSignal: (PointerSignalEvent signal) {
if (signal is PointerScrollEvent) {
print(signal.scrollDelta);
}
},
child: TwoDimensionalGridView(
...
```
Output sample:
```console
flutter: Offset(-7.8, -14.4)
flutter: Offset(-6.1, -14.4)
flutter: Offset(-7.0, -14.4)
flutter: Offset(-6.1, -14.4)
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
Diagonal drag scrolling works as expected:

Diagonal pointer scrolls are not applied on both axes at the same time:

</details>
### Logs
_No response_
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[โ] Flutter (Channel stable, 3.24.5, on Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64, locale
en_US.UTF-8)
โข Flutter version 3.24.5 on channel stable at /home/kamil/.local/apps/flutter
โข Upstream repository https://github.com/flutter/flutter.git
โข Framework revision dec2ee5c1f (7 days ago), 2024-11-13 11:13:06 -0800
โข Engine revision a18df97ca5
โข Dart version 3.5.4
โข DevTools version 2.37.3
[โ] Android toolchain - develop for Android devices
โ Unable to locate Android SDK.
Install Android Studio from: https://developer.android.com/studio/index.html
On first launch it will assist you in installing the Android SDK components.
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
If the Android SDK has been installed to a custom location, please use
`flutter config --android-sdk` to update to that location.
[โ] Chrome - develop for the web
โข CHROME_EXECUTABLE = /usr/bin/brave-browser-stable
[โ] Linux toolchain - develop for Linux desktop
โข clang version 19.1.0 (Fedora 19.1.0-1.fc41)
โข cmake version 3.30.5
โข ninja version 1.12.1
โข pkg-config version 2.3.0
[!] Android Studio (not installed)
โข Android Studio not found; download from https://developer.android.com/studio/index.html
(or visit https://flutter.dev/to/linux-android-setup for detailed instructions).
[โ] VS Code (version 1.95.3)
โข VS Code at /usr/share/code
โข Flutter extension version 3.100.0
[โ] Connected device (2 available)
โข Linux (desktop) โข linux โข linux-x64 โข Fedora Linux 41 (Workstation Edition) 6.11.7-300.fc41.x86_64
โข Chrome (web) โข chrome โข web-javascript โข Brave Browser 131.1.73.89
[โ] Network resources
โข All expected network resources are available.
! Doctor found issues in 2 categories.
```
</details>
|
framework,f: scrolling,platform-web,a: desktop,P2,team-framework,triaged-framework
|
low
|
Minor
|
2,675,698,485 |
vscode
|
Presentation of experimental settings in settings.json
|

Thoughts:
1. Why the line about no quick fixes?
2. I was expecting it to look more like deprecated, where there's a yellow underline to be much more obvious about settings that could potentially break the editor.
|
bug,settings-editor
|
low
|
Minor
|
2,675,717,701 |
vscode
|
Light bulb shows up for move to new file?
|
Maybe I remember wrong but didn't we hide the lightbulb when the only actions were move to file?

Version: 1.96.0-insider (user setup)
Commit: 8d1d1373cd9b8408222b51958e9688827a9e27ce
Date: 2024-11-20T09:23:45.992Z
Electron: 32.2.5
ElectronBuildId: 10579404
Chromium: 128.0.6613.186
Node.js: 20.18.0
V8: 12.8.374.38-electron.0
OS: Windows_NT x64 10.0.26100
|
bug,editor-code-actions
|
low
|
Minor
|
2,675,759,556 |
kubernetes
|
Cronjobs occasionally run much later than scheduled
|
### What happened?
CronJobs will occasionally fail to create their Jobs at the scheduled time, with delays ranging from 3 minutes to over 50 minutes. In the most extreme cases we've observed delays over 60 minutes.
From the kube-controller-manager logs, if I'm understanding correctly (which i may not), it appears that CronJobs rely exclusively on the reconciliation process to trigger Job creation. If no reconciliation occurs around the scheduled time, the Job is not created until a subsequent reconciliation cycle is triggered.
Logs consistently show the following pattern (verbosity set to 8), even for CronJobs that execute on time, this job ran 5 minutes later than scheduled:
```console
I1120 11:05:32.562672 1 cronjob_controllerv2.go:526] "No unmet start times" logger="cronjob-controller" cronjob="camel-test/mkt-cloud-segments-servicecloud-v2-simu-buidar-pp"
I1120 11:05:32.562702 1 cronjob_controllerv2.go:219] "Re-queuing cronjob" logger="cronjob-controller" cronjob="camel-test/mkt-cloud-segments-servicecloud-v2-simu-buidar-pp" requeueAfter="23h54m27.538988861s"
```
Ocasionally this pattern can also be observed in the kube-controller-manager logs:
```console
E1120 13:33:05.831607 1 cronjob_controllerv2.go:168] error syncing CronJobController golang/datafeeds-download-datafeeds, requeuing: Operation cannot be fulfilled on cronjobs.batch "datafeeds-download-datafeeds": the object has been modified; please apply your changes to the latest version and try again
```
Performing manual actions, such as editing the CronJob object or deleting an old Job, forces an immediate reconciliation and triggers the creation of the pending Job.
Another thing we observed is that the issue only affects cronjobs that run on a daily basis (these are distributed at different times) but those that run frequently, like every 15 minutes, never get delayed.
Things we checked:
1. Control-plane clocks verified: All clocks are synchronized using NTP. No drift observed.
2. Increased logging verbosity: Enabled v=8 for kube-controller-manager
3. The number of cronjobs does not seem to be too much for the size of the cluster and resource utilization
We don't know what could be causing this behavior and we don't know how to continue debugging it further.
### What did you expect to happen?
CronJobs should create their corresponding Jobs at the scheduled time.
### How can we reproduce it (as minimally and precisely as possible)?
These are the details of the basic current cluster configuration, however I do not know if the issue could be replicated in another similar cluster.
- Cluster Size: 8 worker nodes, 3 control-plane nodes
- CronJob API Version: batch/v1
- Number of CronJobs: ~100 scheduled at various intervals, distributed in multiple namespaces
Resource Utilization:
- Control-plane nodes have sufficient CPU and memory resources (not near limits).
- The kube-controller-manager pod does not have resource limits set.
### Anything else we need to know?
Keda operator is installed and we have 3 deployments that are managed by Keda, logs regarding these objects are frequent:
```console
11:46:17.755947 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=0 metric="external metric s0-cron-Europe-Madrid-07xx1-5-021xx1-5(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-activemq-dev,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:17 +0100 CET" scaleTarget="Deployment/activemq-test/activemq-dev"
11:46:17.760043 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=1 metric="external metric s0-cron-Europe-Madrid-0000xxx-5523xxx(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-integracions-erpex,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:17 +0100 CET" scaleTarget="Deployment/camel/integracions-erpex"
11:46:31.385874 graph_builder.go:681] "GraphBuilder process object" logger="garbage-collector-controller" apiVersion="keda.sh/v1alpha1" kind="ScaledObject" object="camel/cron-integracions-erpex" uid="d9d7f22a-0eef-4c21-8c23-7823fc6e1f60" eventType="update" virtual=false
11:46:31.900165 graph_builder.go:681] "GraphBuilder process object" logger="garbage-collector-controller" apiVersion="keda.sh/v1alpha1" kind="ScaledObject" object="activemq-test/cron-activemq-dev" uid="eef2ed20-c9ef-40a2-bee4-3c0fcc71dd5a" eventType="update" virtual=false
11:46:32.782034 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=0 metric="external metric s0-cron-Europe-Madrid-07xx1-5-021xx1-5(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-activemq-dev,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:32 +0100 CET" scaleTarget="Deployment/activemq-test/activemq-dev"
11:46:32.787861 horizontal.go:841] "Proposing desired replicas" logger="horizontal-pod-autoscaler-controller" desiredReplicas=1 metric="external metric s0-cron-Europe-Madrid-0000xxx-5523xxx(&LabelSelector{MatchLabels:map[string[]string{scaledobject.keda.sh/name: cron-integracions-erpex,},MatchExpressions:[]LabelSelectorRequirement{},})" timestamp="2024-11-19 11:46:32 +0100 CET" scaleTarget="Deployment/camel/integracions-erpex"
```
There is a custom operator that observes events on several custom resources.
We've observed this issue through multiple kubernetes versions, no upgrade has fixed it.
### Kubernetes version
<details>
```console
$ kubectl version
Client Version: v1.30.6
Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3
Server Version: v1.30.6
```
</details>
### Cloud provider
<details>
On premise
</details>
### OS version
<details>
```console
# On Linux:
PRETTY_NAME="Ubuntu 22.04.5 LTS"
NAME="Ubuntu"
VERSION_ID="22.04"
VERSION="22.04.5 LTS (Jammy Jellyfish)"
VERSION_CODENAME=jammy
ID=ubuntu
ID_LIKE=debian
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
UBUNTU_CODENAME=jammy
$ uname -a
Linux k8scp02 5.15.0-119-generic #129-Ubuntu SMP Fri Aug 2 19:25:20 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
```
</details>
### Install tools
<details>
kubeadm
</details>
### Container runtime (CRI) and version (if applicable)
<details>
containerd 1.7.12-0ubuntu2~22.04.1
</details>
### Related plugins (CNI, CSI, ...) and versions (if applicable)
<details>
vwmare-csi v3.3.1
cilium v1.16.1
</details>
|
kind/bug,sig/apps,needs-triage
|
low
|
Critical
|
2,675,801,028 |
go
|
os/signal: TestSignalTrace is flaky on openbsd/arm64
|
The `os/signal.TestSignalTrace` is flaky on openbsd/arm64:
https://build.golang.org/log/5a334b71cd825842917d59f362ca9c211da82ec9
https://build.golang.org/log/2c4c9b74d2a13ab71e0a00d8430fbf367a4a96e1
This appears to be a signal being lost - it's not currently clear when this issue started, or where the problem exists (and it may well be a kernel related issue). The failure can be reproduced by running `os/signal.TestSignalTrace` in a loop on an Apple M1 or Apple M2. Running another competing workload (such as a build) will speed up the test failure.
This test should be skipped until it can be diagnosed further.
|
OS-OpenBSD,NeedsInvestigation,arch-arm64,compiler/runtime
|
low
|
Critical
|
2,675,889,804 |
ui
|
[bug]: component at https://ui.shadcn.com/r/colors/Gray.json was not found.
|
### Describe the bug
The component at https://ui.shadcn.com/r/colors/Gray.json was not found.
It may not exist at the registry. Please make sure it is a valid component.
### Affected component/components
Drawer
### How to reproduce
1. Go to Docs on Shadcn
2. Click on Drawer
3. follow the installation steps
### Codesandbox/StackBlitz link
_No response_
### Logs
_No response_
### System Info
```bash
Terminal
```
### Before submitting
- [X] I've made research efforts and searched the documentation
- [X] I've searched for existing issues
|
bug
|
low
|
Critical
|
2,675,916,345 |
angular
|
adev: Dropdown design issue on Firefox
|
### Describe the problem that you experienced
Not sure if all dropdowns are affected, but the dropdown on the [Angular Update page](https://angular.dev/update-guide) has two issues on Firefox:
- There is some horizontal scrolling (which starts all the way right, making it impossible to read what it says without scrolling left).
- The scrollbar background is translucent.
### Enter the URL of the topic with the problem
https://angular.dev/update-guide
### Describe what you were looking for in the documentation
_No response_
### Describe the actions that led you to experience the problem
_No response_
### Describe what you want to experience that would fix the problem
_No response_
### Add a screenshot if that helps illustrate the problem
Firefox:

Chrome:

### If this problem caused an exception or error, please paste it here
```true
```
### If the problem is browser-specific, please specify the device, OS, browser, and version
```true
Firefox 132 (on Windows 10, not sure if relevant)
```
### Provide any additional information here in as much as detail as you can
```true
```
|
area: docs-infra
|
low
|
Critical
|
2,675,939,472 |
pytorch
|
[FlexAttention] Block sparse mask is significantly slower than sdpa and handmade triton implementation
|
### ๐ Describe the bug
When testing flex attention with the mask from issue #141126, I found that apart from being incorrect it is also significantly slower than torch's generic `scaled_dot_product_attention` in the example (at lower seq lens < 1000).

Also, it's way slower than handmade (and correct) triton implementation of the same attention.
Ask me if any additional info is needed.
### Versions
PyTorch version: 2.5.0
Is debug build: False
CUDA used to build PyTorch: 12.4
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: version 3.22.1
Libc version: glibc-2.35
Python version: 3.11.10 (main, Sep 7 2024, 18:35:41) [GCC 11.4.0] (64-bit runtime)
Python platform: Linux-5.4.210-39.1-x86_64-with-glibc2.35
Is CUDA available: True
CUDA runtime version: Could not collect
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration:
GPU 0: NVIDIA A800-SXM4-80GB
GPU 1: NVIDIA A800-SXM4-80GB
GPU 2: NVIDIA A800-SXM4-80GB
GPU 3: NVIDIA A800-SXM4-80GB
GPU 4: NVIDIA A800-SXM4-80GB
GPU 5: NVIDIA A800-SXM4-80GB
GPU 6: NVIDIA A800-SXM4-80GB
GPU 7: NVIDIA A800-SXM4-80GB
Nvidia driver version: 550.54.14
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_adv.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_cnn.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_engines_precompiled.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_engines_runtime_compiled.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_graph.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_heuristic.so.9.1.0
/usr/lib/x86_64-linux-gnu/libcudnn_ops.so.9.1.0
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: False
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 48 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 120
On-line CPU(s) list: 0-119
Vendor ID: AuthenticAMD
Model name: AMD EPYC 7662 64-Core Processor
CPU family: 23
Model: 49
Thread(s) per core: 1
Core(s) per socket: 1
Socket(s): 120
Stepping: 0
BogoMIPS: 3992.18
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities
Virtualization: AMD-V
L1d cache: 7.5 MiB (120 instances)
L1i cache: 7.5 MiB (120 instances)
L2 cache: 60 MiB (120 instances)
L3 cache: 1.9 GiB (120 instances)
NUMA node(s): 4
NUMA node0 CPU(s): 0-29
NUMA node1 CPU(s): 30-59
NUMA node2 CPU(s): 60-89
NUMA node3 CPU(s): 90-119
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP disabled, RSB filling, PBRSB-eIBRS Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.24.3
[pip3] torch==2.5.0
[pip3] torch-tb-profiler==0.4.3
[pip3] triton==3.1.0
[conda] Could not collect
cc @msaroufim @jbschlosser @bhosmer @cpuhrsch @erichan1 @drisspg @mikaylagawarecki @Chillee @yanboliang @BoyuanFeng @chauhang @penguinwu @zou3519 @ydwu4 @bdhirsh @yf225
|
module: performance,triaged,module: flex attention,module: sdpa
|
low
|
Critical
|
2,675,957,062 |
material-ui
|
[docs] Convert the "all components" images to CSS
|
### Summary
The [Material UI](https://mui.com/material-ui/all-components/), [Base UI](https://mui.com/base-ui/all-components/), and [Charts](https://mui.com/x/react-charts/#all-mui-x-charts-components) docs have an "all components" page with illustrated cards for each component. These illustrations are currently done on Figma and added to the cards as png. That's not entirely ideal because:
1. Maintaining them on Figma is an additional tool to manage
2. Compressing images and then needing to download them is cumbersome and not efficient performance-wise
3. We can't add micro animations
So, the idea is to, at some point, tackle all of these points by converting them into CSS.
This issue tracks the effort for the Base UI images.
**Search keywords**: component images css
**Search keywords**:
|
docs,design
|
low
|
Major
|
2,675,961,578 |
flutter
|
Build linux failed by INSTALL not found "native_assets/linux"
|
### Steps to reproduce
'flutter build linux' will trigger this bug. I've tried 'flutter clean' or create a sample project but this bug still occurred.
### Expected results
build successfully just as 'flutter build apk' does.
### Actual results
```
CMake Error at cmake_install.cmake:112 (file):
file INSTALL cannot find "/home/....../build/native_assets/linux": No such
file or directory.
```
### Code sample
<details open><summary>Code sample</summary>
```dart
[Paste your code here]
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
[Upload media here]
</details>
### Logs
<details open><summary>Logs</summary>
```console
[Paste your logs here]
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
Doctor summary (to see all details, run flutter doctor -v):
[โ] Flutter (Channel stable, 3.24.5, on Arch Linux 6.11.9-zen1-1-zen, locale en_US.UTF-8)
[!] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
! Some Android licenses not accepted. To resolve this, run: flutter doctor --android-licenses
[โ] Chrome - develop for the web
[โ] Linux toolchain - develop for Linux desktop
[!] Android Studio (not installed)
[โ] IntelliJ IDEA Ultimate Edition (version 2024.3)
[โ] Connected device (2 available)
[โ] Network resources
```
</details>
|
platform-linux,a: desktop,P2,team-linux,triaged-linux
|
low
|
Critical
|
2,675,966,815 |
material-ui
|
[Autocomplete][docs] The Tailwind code demo doesn't include item focus styles
|
### Related page
https://mui.com/base-ui/react-autocomplete/
### Kind of issue
Missing information
### Issue description
The Tailwind autocomplete demo has visible focus on selected items, i.e. if you hover an item or used the up/down keys, items are visibly focused.
However, if you copy the code from the code sample into a project, or if you open the [linked codesandbox](https://codesandbox.io/s/myqs2l?file=/src/Demo.tsx), those visible focus states don't work and there's no visible feedback about which item is focused.
From the devtools, it looks like there are some extra css rules applied to the in-page demo that aren't in the code sample.
### Context
I'm trying to use the base-ui autocomplete component with Tailwind in the way demonstrated on the page.
**Search keywords**: tailwind focus autocomplete base-ui
**Search keywords**:
|
bug ๐,docs,component: autocomplete
|
low
|
Major
|
2,675,978,447 |
ant-design
|
[BUG] Select with search requires double enter hit to trigger the onChange event
|
### Reproduction link
[](https://stackblitz.com/edit/antd-reproduce-5x-qe3zcq?file=demo.tsx)
### Steps to reproduce
1. Choose 'Edoardo' user shown into the Select by click
2. Re-open the Select and fill the text with 'Brugola', then hit 'Enter' (one hit the search works)
3. Re-open the Select and fill the text with 'Pistoni', then hit 'Enter' (not work at the first hit. Hit 'Enter' again and it works)
### What is expected?
Hit the 'Enter' button only once time.
### What is actually happening?
After the first search by text (that works as expected), the Select requires two hit by 'Enter' button to perform the onChange event.
| Environment | Info |
| --- | --- |
| antd | 5.22.1 |
| React | 18.3.1 |
| System | nodeJS latest |
| Browser | Chrome latest |
---
This is happened with React state and REDUX also.
<!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
|
๐ Bug,Inactive
|
low
|
Critical
|
2,675,981,770 |
vscode
|
Horizontal scroll position resets in file when seek peek definition called
|
<!-- โ ๏ธโ ๏ธ Do Not Delete This! bug_report_template โ ๏ธโ ๏ธ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- ๐ฎ Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- ๐ Search existing issues to avoid creating duplicates. -->
<!-- ๐งช Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- ๐ก Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- ๐ง Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- ๐ช If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- ๐ฃ Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.95.2
- OS Version: linux mint mate 21.3
Version: 1.95.2
Commit: e8653663e8840adaf45af01eab5c627a5af81807
Date: 2024-11-07T11:07:22.054Z
Electron: 32.2.1
ElectronBuildId: 10427718
Chromium: 128.0.6613.186
Node.js: 20.18.0
V8: 12.8.374.38-electron.0
OS: Linux x64 5.15.0-126-generic
__Issue:__
With long lines in a code file and horizontally scrolled to bring the remaining part of the line into view
In this scroll position, when tried to seek peek definition
the seek peek definition shows inline,
but the original code file horizontally scrolls back to start position
Steps to Reproduce:
1. Identify long code line in a file and horizontally move to bring a symbol to view
2. select seek peek definition for this symbol
sample images -


( note the horizontal scroll position has come to start position )
and have scroll back to see symbol in code.
__Expected behaviour:__
horizontal scroll position to be retained during seek peek definition
as in sample image below

|
feature-request,ux,polish,editor-symbols
|
low
|
Critical
|
2,675,982,410 |
rustdesk
|
Keyboard German to German Raspberry PI4 not working
|
### Bug Description
Hello,
If i connect with Rustdesk to a Raspberry PI4 local Language German:

The Keys are not forwarded correctly. For example instead of * i get }. Do you have a Idea how to fix this?
Thanks
### How to Reproduce
Install Rustdesk under Raspbian and connect.
### Expected Behavior
Correct keys.
### Operating system(s) on local (controlling) side and remote (controlled) side
Windows 10 -> Raspbian
### RustDesk Version(s) on local (controlling) side and remote (controlled) side
1.3.1 - > 1.2.7
### Screenshots

### Additional Context
_No response_
|
bug
|
low
|
Critical
|
2,675,996,407 |
vscode
|
Enable highlighting scroll in copilot chat window
|
<!-- โ ๏ธโ ๏ธ Do Not Delete This! feature_request_template โ ๏ธโ ๏ธ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
In the Github Copilot chat window it can sometimes give very long answers, rewriting entire files/scripts, when only a portion of that code is actually changing. Currently, when I try to highlight a longer section of code but don't want all of it, I need to click the start of where I want to copy, scroll down manually, and click the end with shift held to highlight everything. Instead it would be easier if I could highlight and have the scrollbar move with the highlighted section as it hits the top/bottom of the currently visible window. This type of feature is implemented in most other programs with scroll text boxes.
|
chat
|
low
|
Minor
|
2,676,046,424 |
node
|
FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal
|
### Version
v22.11.0
### Platform
```text
Darwin 23.6.0 Darwin Kernel Version 23.6.0
I also tested it on Linux machine.
```
### Subsystem
_No response_
### What steps will reproduce the bug?
Hi,
I want to report a node bug. It can reproduced by running the following PoC, node.js gives a FATAL error.
```javascript
const {spawn} = require('child_process');
Object.defineProperty(Array.prototype, "0", {
set(){
console.log(123);
}
});
spawn("ls");
```
Please note that this bug differs from the reported [here](https://github.com/nodejs/node/issues/54186).
AH.
### How often does it reproduce? Is there a required condition?
By just running the given code, node runtime gives node crash.
### What is the expected behavior? Why is that the expected behavior?
Not a crash, by looking at the stack trace, it is connected to `v8::ToLocalChecked Empty MaybeLocal`.
### What do you see instead?
FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal
FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal
----- Native stack trace -----
1: 0x10bdd2277 node::OnFatalError(char const*, char const*)
2: 0x10c025171 v8::Utils::ReportApiFailure(char const*, char const*)
3: 0x10bef84cc node::(anonymous namespace)::ProcessWrap::Spawn(v8::FunctionCallbackInfo<v8::Value> const&)
4: 0x10cb21a62 Builtins_CallApiCallbackGeneric [~/node/v22.11.0/bin/node]
----- JavaScript stack trace -----
1: ChildProcess.spawn (node:internal/child_process:395:28)
2: spawn (node:child_process:759:9)
3: ${PATH_TO_POC}
4: node:internal/modules/cjs/loader:1546:14
5: node:internal/modules/cjs/loader:1689:10
6: node:internal/modules/cjs/loader:1318:32
7: node:internal/modules/cjs/loader:1128:12
8: traceSync (node:diagnostics_channel:315:14)
9: wrapModuleLoad (node:internal/modules/cjs/loader:218:24)
10: executeUserEntryPoint (node:internal/modules/run_main:170:5)
### Additional information
_No response_
|
confirmed-bug,child_process
|
low
|
Critical
|
2,676,131,157 |
ollama
|
High Inference Time and Limited GPU Utilization with Ollama Docker
|
### What is the issue?
## Description:
I am using Ollama in a Docker setup with GPU support, configured to use all available GPUs on my system. However, when using the NemoTron model with a simple prompt and utilizing the function calling feature, the inference time is around 50 seconds to get a response, which is too high for my use case.
## Docker Configuration:
Here is my docker-compose.yml file for Ollama:
services:
ollama:
image: ollama/ollama
container_name: ollama
hostname: ollama
ports:
- "11434:11434"
volumes:
- /home/system/dockers/volumes/ollama:/root/.ollama
runtime: nvidia
environment:
- NVIDIA_VISIBLE_DEVICES=all
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: all
capabilities: [gpu]
restart: unless-stopped
networks:
- genai_network
networks:
genai_network:
## Log Details:
### Server Initialization:
Listening on [::]:11434 (version 0.4.2)
Dynamic LLM libraries runners="[cpu_avx2 cuda_v11 cuda_v12 cpu cpu_avx]"
Looking for compatible GPUs
Inference compute id=GPU-660ca8b7-181a-ede9-f6fe-8ccd5f9dbb89 library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="47.1 GiB"
Inference compute id=GPU-d62f5e11-4192-0e70-0732-55b558edcb7a library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="46.4 GiB"
Inference compute id=GPU-45972459-815c-f304-9fdf-b952276c9b13 library=cuda variant=v12 compute=8.9 driver=12.4 name="NVIDIA RTX 6000 Ada Generation" total="47.5 GiB" available="47.0 GiB"
### Model Loading:
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA RTX 6000 Ada Generation, compute capability 8.9, VMM: yes
llm_load_tensors: ggml ctx size = 0.68 MiB
## Observed Issue:
During inference with the NemoTron model, the response time is around 50 seconds.
The logs show that only one GPU seems to be utilized (found 1 CUDA devices), despite multiple GPUs being detected during service initialization.
## Questions:
### GPU Load Balancing:
Does Ollama support load balancing across multiple GPUs? If yes, why do the logs indicate that only one GPU is used (found 1 CUDA devices) when the model is loaded?
### Performance Optimization:
What steps are recommended to reduce inference time?
Should I adjust configuration settings in Docker or Ollama?
Could variables like GGML_CUDA_FORCE_CUBLAS or GGML_CUDA_FORCE_MMQ improve performance?
## Technical Context:
Ollama Version: 0.4.2
Hardware Configuration:
GPUs: 3 x NVIDIA RTX 6000 Ada Generation (47.5 GiB VRAM each)
NVIDIA Driver: 12.4
Model Used: NemoTron
Usage Scenario: Simple prompt with function calling.
## Expectation:
- Confirmation on multi-GPU support in Ollama.
- Suggestions to reduce inference time.
- Documentation or examples of optimized configuration for heavy workloads with multiple GPUs.
## Thank You for Your Help!

### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.4.2
|
bug
|
low
|
Major
|
2,676,135,808 |
flutter
|
[go_router] Browser back button unexpectedly navigates to an already replaced route that was the result of a redirection
|
### Steps to reproduce
1. Launch the sample
2. Tap the "Go to child screen" button
3. Wait 2 seconds for the deep link screen to forward you to the child screen
4. Being on the child screen, tap the browser's back button
### Expected results
You see the home screen
### Actual results
You see the deep link screen
### Code sample
Link to Dartpad
https://dartpad.dev/?id=72e8664b6408534f2448cc8a34199ee6
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/foundation.dart';
import 'package:flutter/material.dart';
import 'package:go_router/go_router.dart';
// 1. The main file
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
@override
Widget build(BuildContext context) {
return MaterialApp.router(routerConfig: goRouter);
}
}
// 2. The GoRouter configuration
final goRouter = GoRouter(
debugLogDiagnostics: kDebugMode,
routes: [
GoRoute(
path: '/',
builder: (context, state) => const SplashScreen(),
),
GoRoute(
path: '/home',
builder: (context, state) => const HomeScreen(),
),
GoRoute(
path: '/deeplink',
builder: (context, state) => const DeepLinkScreen(),
),
GoRoute(
path: '/child',
builder: (context, state) => const ChildScreen(),
redirect: (context, GoRouterState state) {
final valueAddedDuringDeepLink = state.uri.queryParameters['value'];
if (valueAddedDuringDeepLink == '42') {
return null;
} else {
return '/deeplink';
}
},
),
],
);
// 3. The 4 screens
// 3.1 Splash screen
class SplashScreen extends StatefulWidget {
const SplashScreen({super.key});
@override
State<SplashScreen> createState() => _SplashScreenState();
}
class _SplashScreenState extends State<SplashScreen> {
@override
void initState() {
super.initState();
Future.delayed(const Duration(seconds: 1), () {
context.go('/home');
});
}
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(title: const Text('Splash screen')),
body: const Center(
child: Column(
mainAxisSize: MainAxisSize.min,
children: [
Text('Loading...'),
CircularProgressIndicator(),
],
)),
);
}
}
// 3.2 Home screen
class HomeScreen extends StatelessWidget {
const HomeScreen({super.key});
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(title: const Text('Home screen')),
body: Center(
child: Column(
mainAxisSize: MainAxisSize.min,
children: [
const Text('Home'),
TextButton(
onPressed: () => context.push('/child'),
child: const Text('Go to child screen'),
),
],
)),
);
}
}
// 3.3 Deep link screen
class DeepLinkScreen extends StatefulWidget {
const DeepLinkScreen({super.key});
@override
State<DeepLinkScreen> createState() => _DeepLinkScreenState();
}
class _DeepLinkScreenState extends State<DeepLinkScreen> {
@override
void initState() {
super.initState();
Future.delayed(const Duration(seconds: 1), () {
context.pushReplacement('/child?value=42');
});
}
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(title: const Text('Deep link screen')),
body: const Center(
child: Column(
mainAxisSize: MainAxisSize.min,
children: [
Text('Redirecting to the child screen...'),
CircularProgressIndicator(),
],
)));
}
}
// 3.4 Child screen
class ChildScreen extends StatelessWidget {
const ChildScreen({super.key});
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(title: const Text('Child screen')),
body: const Center(child: Text('Child')),
);
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
https://github.com/user-attachments/assets/fcfe8c3e-bbe6-4747-b864-0039a2e858c9
</details>
### Logs
<details open><summary>Logs</summary>
```console
Launching lib\main.dart on Chrome in debug mode...
Waiting for connection from debug service on Chrome...
This app is linked to the debug service: ws://127.0.0.1:63040/Dkr73jPlCtM=/ws
Debug service listening on ws://127.0.0.1:63040/Dkr73jPlCtM=/ws
Debug service listening on ws://127.0.0.1:63040/Dkr73jPlCtM=/ws
[GoRouter] Full paths for routes:
โโ/ (SplashScreen)
โโ/home (HomeScreen)
โโ/deeplink (DeepLinkScreen)
โโ/child (ChildScreen)
[GoRouter] setting initial location null
[GoRouter] Using MaterialApp configuration
[GoRouter] going to /home
[GoRouter] pushing /child
[GoRouter] redirecting to RouteMatchList#7876f(uri: /deeplink, matches: [RouteMatch#f0590(route: GoRoute#6038c(name: null, path: "/deeplink"))])
[GoRouter] pushReplacement /child?value=42
[GoRouter] pushing /child
[GoRouter] redirecting to RouteMatchList#7876f(uri: /deeplink, matches: [RouteMatch#f0590(route: GoRoute#6038c(name: null, path: "/deeplink"))])
[GoRouter] pushReplacement /child?value=42
[GoRouter] pushReplacement /child?value=42
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
PS G:\Projects\gorouter-playground> flutter doctor -v
[โ] Flutter (Channel stable, 3.19.4, on Microsoft Windows [Version 10.0.22631.4460], locale en-US)
โข Flutter version 3.19.4 on channel stable at C:\Development\SDKs\Flutter
โข Upstream repository https://github.com/flutter/flutter.git
โข Framework revision 68bfaea224 (8 months ago), 2024-03-20 15:36:31 -0700
โข Engine revision a5c24f538d
โข Dart version 3.3.2
โข DevTools version 2.31.1
[โ] Windows Version (Installed version of Windows is version 10 or higher)
[โ] Android toolchain - develop for Android devices (Android SDK version 34.0.0)
โข Android SDK at C:\Development\SDKs\Android
โข Platform android-34, build-tools 34.0.0
โข ANDROID_HOME = C:\Development\SDKs\Android
โข ANDROID_SDK_ROOT = C:\Development\SDKs\Android
โข Java binary at: C:\Development\IDEs\AndroidStudioKoala\jbr\bin\java
โข Java version OpenJDK Runtime Environment (build 17.0.11+0--11852314)
โข All Android licenses accepted.
[โ] Chrome - develop for the web
โข Chrome at C:\Program Files\Google\Chrome\Application\chrome.exe
[โ] Visual Studio - develop Windows apps (Visual Studio Community 2022 17.11.6)
โข Visual Studio at C:\Program Files\Microsoft Visual Studio\2022\Community
โข Visual Studio Community 2022 version 17.11.35431.28
โข Windows 10 SDK version 10.0.22621.0
[โ] Android Studio (version 2024.1)
โข Android Studio at C:\Development\IDEs\AndroidStudioKoala
โข Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
โข Dart plugin can be installed from:
https://plugins.jetbrains.com/plugin/6351-dart
โข Java version OpenJDK Runtime Environment (build 17.0.11+0--11852314)
[โ] IntelliJ IDEA Community Edition (version 2022.1)
โข IntelliJ at C:\Development\IDEs\IntelliJ IDEA\IntelliJ IDEA Community Edition 2020.3.2
โข Flutter plugin can be installed from:
https://plugins.jetbrains.com/plugin/9212-flutter
โข Dart plugin version 221.6103.1
[โ] Connected device (3 available)
โข Windows (desktop) โข windows โข windows-x64 โข Microsoft Windows [Version 10.0.22631.4460]
โข Chrome (web) โข chrome โข web-javascript โข Google Chrome 131.0.6778.70
โข Edge (web) โข edge โข web-javascript โข Microsoft Edge 130.0.2849.68
[โ] Network resources
โข All expected network resources are available.
โข No issues found!
```
</details>
|
platform-web,package,has reproducible steps,P2,p: go_router,team-go_router,triaged-go_router,found in release: 3.24,found in release: 3.27
|
low
|
Critical
|
2,676,138,749 |
TypeScript
|
MIDIMessageEvent data is incorrectly typed
|
### ๐ Search Terms
In https://github.com/microsoft/TypeScript/pull/57027 `MIDIMessageEvent.data` was changed from `Uint8Array` to `Uint8Array | null` (see https://github.com/microsoft/TypeScript/pull/57027/files#diff-dc0eab937d15e62545da3ed7b4f40ad6b24f15dd88fbc6ceda2bfb4ed8356eb0L14670).
According to https://developer.mozilla.org/en-US/docs/Web/API/MIDIMessageEvent/data data the value will always be an Uint8Array.
### ๐ Version & Regression Information
This changed in #57027
### โฏ Playground Link
_No response_
### ๐ป Code
### ๐ Actual behavior
`'event.data' is possibly 'null'.ts(18047)`
### ๐ Expected behavior
No warning about data property possibly being null.
### Additional information about the issue
_No response_
|
Bug,Help Wanted,Domain: lib.d.ts
|
low
|
Minor
|
2,676,203,778 |
react
|
[React 19] useOptimistic - Add option to disable rollback
|
## Summary
`useOptimistic` automatically rolls back upon error. Can we make this optional?
Use case: I would like to keep the optimistic data displaying upon error, and display a "Try again" button next to it. (yes, I could do this without useOptimistic, but I suspect this will be a common use case).
Also, the current rollback on error behavior [isn't documented here](https://react.dev/reference/react/useOptimistic), so suggest adding before release.
<!--
Please provide a CodeSandbox (https://codesandbox.io/s/new), a link to a
repository on GitHub, or provide a minimal code example that reproduces the
problem. You may provide a screenshot of the application if you think it is
relevant to your bug report. Here are some tips for providing a minimal
example: https://stackoverflow.com/help/mcve.
-->
|
React 19
|
medium
|
Critical
|
2,676,240,652 |
godot
|
Performance drop in 3D scenes with DirectionalLight in Godot 3.6 compared to 3.5
|
### Tested versions
- Godot 3.5.3
- Godot 3.6
### System information
Linux Ubuntu 22.04, Intel Core 2 Duo 2.8 Ghz, Geforce 9500 GT 3Gb VRAM, 4 Gb RAM 800 Mhz
### Issue description
Performance is considerably worse in Godot 3.6, compared to 3.5
Godot 3.5.3, with 33 FPS:

Same exact scene in Godot 3.6 with 27 FPS:

A drop of 20% in performance can be observed.
### Steps to reproduce
Run a project in Godot 3.5.3, make note of the FPS
Run the same project in Godot 3.6 and observe the drop in performance.
### Minimal reproduction project (MRP)
The truck town demo can be used to attest the performance change.
https://github.com/godotengine/godot-demo-projects/releases/download/3.5-9e68af3/3d_truck_town.zip
|
bug,topic:rendering,regression,topic:3d,performance
|
low
|
Major
|
2,676,249,674 |
ollama
|
Ollama on AMD CPU iGPU Windows Hack
|
Followed this tutorial https://github.com/ollama/ollama/pull/5059
and got Ollama running on windows on AMD CPU,
any tips on how I get it to use iGPU what commands can i try to hack?
Thanks
|
feature request
|
low
|
Minor
|
2,676,290,606 |
rust
|
Tracking Issue for `vec_deque_iter_as_slices`
|
<!--
Thank you for creating a tracking issue!
Tracking issues are for tracking a feature from implementation to stabilization.
Make sure to include the relevant RFC for the feature if it has one.
If the new feature is small, it may be fine to skip the RFC process. In that
case, you can use `issue = "none"` in your initial implementation PR. The
reviewer will ask you to open a tracking issue if they agree your feature can be
added without an RFC.
-->
Feature gate: `#![feature(vec_deque_iter_as_slices)]`
This is a tracking issue for:
- `alloc::collections::vec_deque::Iter::as_slices`
- `alloc::collections::vec_deque::IterMut::into_slices`
- `alloc::collections::vec_deque::IterMut::as_slices`
- `alloc::collections::vec_deque::IterMut::as_mut_slices`
<!--
Include a short description of the feature.
-->
Each of these functions obtains a pair of slices from a `VecDeque` iterator, similar to `VecDeque::as_slices`.
### Public API
<!--
For most library features, it'd be useful to include a summarized version of the public API.
(E.g. just the public function signatures without their doc comments or implementation.)
-->
```rust
use std::collections::vec_deque::{Iter, IterMut};
impl<'a, T> Iter<'a, T> {
pub fn as_slices(&self) -> (&'a [T], &'a [T]) { ... }
}
impl<'a, T> IterMut<'a, T> {
pub fn into_slices(self) -> (&'a mut [T], &'a mut [T]) { ... }
pub fn as_slices(&self) -> (&[T], &[T]) { ... }
pub fn as_mut_slices(&mut self) -> (&mut [T], &mut [T]) { ... }
}
```
### Steps / History
<!--
For larger features, more steps might be involved.
If the feature is changed later, please add those PRs here as well.
-->
- [x] Implementation: #123947
- [ ] Final comment period (FCP)[^1]
- [ ] Stabilization PR
<!--
Once the feature has gone through a few release cycles and there are no
unresolved questions left, the feature might be ready for stabilization.
If this feature didn't go through the RFC process, a final comment period
(FCP) is always needed before stabilization. This works as follows:
A library API team member can kick off the stabilization process, at which point
the rfcbot will ask all the team members to verify they agree with
stabilization. Once enough members agree and there are no concerns, the final
comment period begins: this issue will be marked as such and will be listed
in the next This Week in Rust newsletter. If no blocking concerns are raised in
that period of 10 days, a stabilization PR can be opened by anyone.
-->
### Unresolved Questions
<!--
Include any open questions that need to be answered before the feature can be
stabilised. If multiple (unrelated) big questions come up, it can be a good idea
to open a separate issue for each, to make it easier to keep track of the
discussions.
It's useful to link any relevant discussions and conclusions (whether on GitHub,
Zulip, or the internals forum) here.
-->
- None yet.
[^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html
|
T-libs-api,C-tracking-issue
|
low
|
Minor
|
2,676,331,380 |
excalidraw
|
Increase text element padding in rectangle
|
As of now, the text padding inside the rectangle is way too small, so it's often necessary to manually add the spaces to achieve a more pleasant look.
Instead, we shall increase the padding, so that no manual "alignment" is not necessary.
Things to keep in mind:
- it might be a regression -> double-check
- do not create a breaking change for existing diagrams (after a manual redraw, like entering wysiwyg, it's likely fine)
- don't forget about padding while editing (wysiwyg)
- test different sizes & zoom levels

|
good first issue,UX/UI
|
low
|
Major
|
2,676,337,729 |
flutter
|
[web] [Pointer Interceptor] App crashed on Chrome when closing a dialog containing Pointer Interceptor and Voice over is turn on
|
### Steps to reproduce
1. Run the app
2. Turn Voice over on
3. Click to open the dialog
4. Close the dialog from the Enable or Disabled button
### Expected results
Navigated back to initial screen
### Actual results
Crash the app on Chrome with Error code 11
### Code sample
<details open><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
import 'package:pointer_interceptor/pointer_interceptor.dart';
Future<void> main() async =>
runApp(const ShowDialogExampleApp());
class ShowDialogExampleApp extends StatelessWidget {
const ShowDialogExampleApp({super.key});
@override
Widget build(BuildContext context) {
return const MaterialApp(
home: DialogExample(),
);
}
}
class DialogExample extends StatelessWidget {
const DialogExample({super.key});
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(title: const Text('showDialog Sample')),
body: Center(
child: OutlinedButton(
onPressed: () => _dialogBuilder(context),
child: const Text('Open Dialog'),
),
),
);
}
Future<void> _dialogBuilder(BuildContext context) {
return showDialog<void>(
context: context,
builder: (BuildContext context) {
return AlertDialog(
title: const Text('Basic dialog title'),
content: const Text(
'A dialog is a type of modal window that\n'
'appears in front of app content to\n'
'provide critical information, or prompt\n'
'for a decision to be made.',
),
actions: <Widget>[
PointerInterceptor(
child: TextButton(
style: TextButton.styleFrom(
textStyle: Theme.of(context).textTheme.labelLarge,
),
child: const Text('Disable'),
onPressed: () {
Navigator.of(context).pop();
},
),
),
TextButton(
style: TextButton.styleFrom(
textStyle: Theme.of(context).textTheme.labelLarge,
),
child: const Text('Enable'),
onPressed: () {
Navigator.of(context).pop();
},
),
],
);
},
);
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Screenshots / Video demonstration</summary>
[Upload media here]
</details>
<img width="1840" alt="Screenshot 2024-11-20 at 17 12 26" src="https://github.com/user-attachments/assets/71b54cd7-5816-49d0-880a-202fc3f8126c">
<img width="1840" alt="Screenshot 2024-11-20 at 17 13 11" src="https://github.com/user-attachments/assets/fbe3ce7a-f5e9-4423-a7f0-9481af2c631f">
<img width="1728" alt="Screenshot 2024-11-20 at 17 13 48" src="https://github.com/user-attachments/assets/c483da20-eea3-492f-b00c-3b1bcf1eb9fe">
### Logs
<details open><summary>Logs</summary>
```console
Debug service listening on ws://127.0.0.1:62102/PNkeCMZovkk=/ws
Debug service listening on ws://127.0.0.1:62102/PNkeCMZovkk=/ws
Debugger: Target crashed!
Debugger: Target crashed!
Debugger: Target crashed!
```
</details>
### Flutter Doctor output
<details open><summary>Doctor output</summary>
```console
[Paste your output here]
```
</details>
|
c: crash,a: accessibility,platform-web,package,has reproducible steps,P2,p: pointer_interceptor,team-web,triaged-web,found in release: 3.27,found in release: 3.28
|
low
|
Critical
|
2,676,379,804 |
tensorflow
|
The documentation in `data_performance.ipynb` uses `py_function()` without an explanation
|
### Issue type
Documentation Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
TF 2.18
### Custom code
No
### OS platform and distribution
_No response_
### Mobile device
_No response_
### Python version
_No response_
### Bazel version
_No response_
### GCC/compiler version
_No response_
### CUDA/cuDNN version
_No response_
### GPU model and memory
_No response_
### Current behavior?
In the ["Better performance with the tf.data API" guide](https://www.tensorflow.org/guide/data_performance), `tf.py_function()` is used several times in the mapping function. There are various performance boosts demonstrated by the guide with said mapping functions. However, there does not seem to be any practical reason why `tf.py_function()` is used inside the mapping functions. In fact, if you remove those, the behavior is the same; in other words, there doesn't seem to be a need for them at all.
Curiously, if you remove them from the examples and then perform the time measurements, the speedup goes away. For example, consider the following mapping function from [the guide](https://github.com/tensorflow/docs/blob/bbc0b9c70fc0bd4411793d1b0bcc56ef1dbc2405/site/en/guide/data_performance.ipynb#L447-L450):
```python
def mapped_function(s):
# Do some hard pre-processing
tf.py_function(lambda: time.sleep(0.03), [], ())
return s
```
In the sequential case, it is shown that the following mapping code results in a time measurement of 0.49222556600034295:
```python
benchmark(
ArtificialDataset()
.map(mapped_function)
)
```
And, the optimized (parallel) version as follows results in a time measurement of 0.36392719900049997:
```python
benchmark(
ArtificialDataset()
.map(
mapped_function,
num_parallel_calls=tf.data.AUTOTUNE
)
)
```
But, if I remove the `tf.py_function()` from the mapping function, I get comparable measurements from both examples, namely, 0.22448736599994845 and 0.2266392660001202:
```python
def mapped_function(s):
# Do some hard pre-processing
lambda: time.sleep(0.03), [], ()
return s
```
In fact, that measurement is even better, which makes me believe that this example is contrived to show a performance benefit by using `num_parallel_calls` when in fact TF is already optimizing the code without it. Frivolously wrapping the function in `tf.py_function()` is most likely causing TensorFlow *not* to optimize the function. Thus, is `num_parallel_calls` even needed to achieve better performance?
### Standalone code to reproduce the issue
```shell
Mentioned in the above description.
```
### Relevant log output
_No response_
|
type:docs-bug,type:bug
|
low
|
Critical
|
2,676,399,871 |
terminal
|
Rendering with tcell under WSL is slow for the first 65s
|
### Windows Terminal version
1.23.3231.0
### Windows build number
10.0.26100.0
### Other Software
_No response_
### Steps to reproduce
Under WSL run this: https://github.com/gdamore/tcell/blob/main/_demos/stress.go
### Expected Behavior
_No response_
### Actual Behavior
Initially the application runs super slow. After 65 seconds it suddenly becomes extremely fast (= the expected performance). Here you can see the transition between the two:
https://github.com/user-attachments/assets/765f5cbe-88b2-4c7e-a21c-bbdefb555687
|
Product-Conpty,Issue-Bug,Product-Terminal,Needs-Tag-Fix,Priority-2
|
low
|
Major
|
2,676,443,002 |
kubernetes
|
Adopting new libraries from cel-go
|
### What would you like to be added?
CEL is general used in Kubernetes now. We wanna keep up with new libraries added into cel-go. cel recently released the [new release](https://github.com/google/cel-go/releases/tag/v0.22.0) which has some feature we love to include in Kubernetes. This issue is to track the adoption of [email protected].
More info: cel introduced versioning for newly added libraries. Kubernetes would love to keep up with supporting the libraries in versions.
### Why is this needed?
/sig api-machinery
/triage accepted
|
sig/api-machinery,kind/feature,triage/accepted
|
low
|
Minor
|
2,676,445,140 |
pytorch
|
CuDNN SDPA Issue Tracker
|
# Summary
Keeps track of known issues and the current state of the integration, with the goal to bump order in priority list with higher confidence
```[tasklist]
### Tasks
- [ ] https://github.com/pytorch/pytorch/issues/139298
- [x] Match Query's stride layout https://github.com/pytorch/pytorch/pull/138354
- [x] gradOutput vs Ouput stride mismatch, blocked by cudnn bump: https://github.com/pytorch/pytorch/pull/138354#issuecomment-2488863424
- [ ] Fill Value twiddiling: https://github.com/pytorch/pytorch/pull/140837
- [ ] Debugging PR: https://github.com/pytorch/pytorch/pull/139699
- [ ] Update CuDNN backend to respect enable_gqa https://github.com/pytorch/pytorch/issues/139586
- [ ] Mitigated: https://github.com/pytorch/pytorch/issues/138529 but should see if fixed in new version of cudnn
- [ ] WSL: https://github.com/huggingface/diffusers/issues/9704 | need to fix or guard correctly
```
cc @csarofeen @ptrblck @xwang233 @eqy @msaroufim @mikaylagawarecki
|
module: cudnn,module: cuda,triaged,module: sdpa
|
low
|
Critical
|
2,676,447,125 |
PowerToys
|
Time Picker for Awake - 12H option
|
### Description of the new feature / enhancement
Can we add option for 24H vs 12H option for picking end time of awake expire?
### Scenario when this would be used?
Choosing end time for Awake expire on specific time and date.
### Supporting information
_No response_
|
Idea-Enhancement,Needs-Triage,Product-Awake
|
low
|
Minor
|
2,676,456,476 |
go
|
testing: panic in AllocsPerRun during parallel test
|
I propose that testing.AllocsPerRun panic if called during a parallel test,
because it cannot give accurate results in that context. Tests that use
the result in that context will be flaky.
For example, flake #70327 was caused by a test that called t.Parallel
and then used AllocsPerRun. The other tests running in parallel
had allocations that were observed by the AllocsPerRun, resulting
in inconsistent results and a flaky failure.
I've implemented this change in [CL 630137](https://go.dev/cl/630137).
|
Proposal,Proposal-Accepted
|
low
|
Critical
|
2,676,482,668 |
kubernetes
|
Migrating cel ext library cost close to the func
|
### What would you like to be added?
Current we keep the extension libraries added in Kubernetes located at: https://github.com/kubernetes/kubernetes/tree/master/staging/src/k8s.io/apiserver/pkg/cel/library
All the cost defined for the ext library is defined in https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apiserver/pkg/cel/library/cost.go which is separately from the definition of func. Future contributors who is not familiar with the ext library would easily missing adding cost while adding new library. It would be great to have cost defined closer to the func.
### Why is this needed?
Good for code maintenance.
/sig api-machinery
/triage accepted
|
sig/api-machinery,kind/feature,triage/accepted
|
low
|
Minor
|
2,676,595,627 |
godot
|
Error when using Move/Duplicate To
|
### Tested versions
Engine version: Godot Engine v4.4.dev4.official (36e6207bb48fafd43b98a458d0c461ad982b4e68)
### System information
Godot v4.4.dev4 - Pop!_OS 22.04 LTS on X11 - X11 display driver, Multi-window, 2 monitors - Vulkan (Forward+) - dedicated Quadro P1000 (nvidia; 560.35.03) - Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz (12 threads)
### Issue description
Godot crashes when trying to copy a file to another location using the Move'/Duplicate To option in the FileSystem.
This is the error I get:
```
ERROR: Task '_update_scan_actions' already exists.
at: add_task (editor/progress_dialog.cpp:173)
================================================================
handle_crash: Program crashed with signal 11
Engine version: Godot Engine v4.4.dev4.official (36e6207bb48fafd43b98a458d0c461ad982b4e68)
Dumping the backtrace. Please include this when reporting the bug to the project developer.
[1] /lib/x86_64-linux-gnu/libc.so.6(+0x42520) [0x7c0f70e42520] (??:0)
[2] /usr/local/bin/godot() [0x1588dd0] (??:0)
[3] /usr/local/bin/godot() [0x158b9c3] (??:0)
[4] /usr/local/bin/godot() [0x4215a28] (??:0)
[5] /usr/local/bin/godot() [0x25ec59c] (??:0)
[6] /usr/local/bin/godot() [0x2602c93] (??:0)
[7] /usr/local/bin/godot() [0x2606159] (??:0)
[8] /usr/local/bin/godot() [0x52ee92] (??:0)
[9] /usr/local/bin/godot() [0x184fe53] (??:0)
[10] /usr/local/bin/godot() [0x16677c1] (??:0)
[11] /usr/local/bin/godot() [0x1589215] (??:0)
[12] /usr/local/bin/godot() [0x158bf78] (??:0)
[13] /usr/local/bin/godot() [0x4264c7d] (??:0)
[14] /usr/local/bin/godot() [0x26060fc] (??:0)
[15] /usr/local/bin/godot() [0x52ee92] (??:0)
[16] /usr/local/bin/godot() [0x41c2fa] (??:0)
[17] /lib/x86_64-linux-gnu/libc.so.6(+0x29d90) [0x7c0f70e29d90] (??:0)
[18] /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0x80) [0x7c0f70e29e40] (??:0)
[19] /usr/local/bin/godot() [0x43af1a] (??:0)
-- END OF BACKTRACE --
================================================================
```
### Steps to reproduce
Right click in one file in the FileSystem section and chose Move/Duplicate To option. Then select another location in the dialog and chose Copy on the options below. Godot crashes completely.
### Minimal reproduction project (MRP)
I'm using an empty 2D project with just a few folders and files, nothing special.
|
bug,topic:editor,needs testing,crash
|
low
|
Critical
|
2,676,609,698 |
bitcoin
|
Functional tests: `feature_bind_port_discover.py` is failing
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current behaviour
The test is failing due to `Discover()` not being able to finding the local addresses "1.1.1.1" and "2.2.2.2".
### Expected behaviour
The test should pass. Addr1 and Addr2 should be found.
### Steps to reproduce
Run the following commands:
```bash
sudo ifconfig lo:0 1.1.1.1/32 up && sudo ifconfig lo:1 2.2.2.2/32 up
./build/test/functional/test_runner.py --ihave1111and2222 feature_bind_port_discover.py
```
### Relevant log output
```bash
andre@acer:~/repos/bitcoin$ build/test/functional/test_runner.py --ihave1111and2222 feature_bind_port_discover.py
Temporary test directory at /tmp/test_runner_โฟ_๐_20241120_141406
Remaining jobs: [feature_bind_port_discover.py]
1/1 - feature_bind_port_discover.py failed, Duration: 1 s
stdout:
2024-11-20T17:14:06.192000Z TestFramework (INFO): PRNG seed is: 4142310399432710311
2024-11-20T17:14:06.193000Z TestFramework (INFO): Initializing test directory /tmp/test_runner_โฟ_๐_20241120_141406/feature_bind_port_discover_0
2024-11-20T17:14:06.511000Z TestFramework (INFO): Test that if -bind= is not passed then all addresses are added to localaddresses
2024-11-20T17:14:06.512000Z TestFramework (ERROR): Assertion failed
Traceback (most recent call last):
File "/home/andre/repos/bitcoin/test/functional/test_framework/test_framework.py", line 132, in main
self.run_test()
File "/home/andre/repos/bitcoin/build/test/functional/feature_bind_port_discover.py", line 63, in run_test
assert found_addr1
AssertionError
2024-11-20T17:14:06.563000Z TestFramework (INFO): Stopping nodes
2024-11-20T17:14:06.716000Z TestFramework (WARNING): Not cleaning up dir /tmp/test_runner_โฟ_๐_20241120_141406/feature_bind_port_discover_0
2024-11-20T17:14:06.716000Z TestFramework (ERROR): Test failed. Test logging available at /tmp/test_runner_โฟ_๐_20241120_141406/feature_bind_port_discover_0/test_framework.log
2024-11-20T17:14:06.716000Z TestFramework (ERROR):
2024-11-20T17:14:06.716000Z TestFramework (ERROR): Hint: Call /home/andre/repos/bitcoin/test/functional/combine_logs.py '/tmp/test_runner_โฟ_๐_20241120_141406/feature_bind_port_discover_0' to consolidate all logs
2024-11-20T17:14:06.716000Z TestFramework (ERROR):
2024-11-20T17:14:06.716000Z TestFramework (ERROR): If this failure happened unexpectedly or intermittently, please file a bug and provide a link or upload of the combined log.
2024-11-20T17:14:06.716000Z TestFramework (ERROR): https://github.com/bitcoin/bitcoin/issues
2024-11-20T17:14:06.716000Z TestFramework (ERROR):
stderr:
TEST | STATUS | DURATION
feature_bind_port_discover.py | โ Failed | 1 s
ALL | โ Failed | 1 s (accumulated)
Runtime: 1 s
```
### How did you obtain Bitcoin Core
Compiled from source
### What version of Bitcoin Core are you using?
Bitcoin Core daemon version v28.99.0-22ef95dbe3e4
### Operating system and version
Ubuntu 22.04
### Machine specifications
_No response_
|
Tests,CI failed
|
low
|
Critical
|
2,676,610,602 |
flutter
|
[Android] Consider implementing backwards compatibility for sensitive content protection
|
Flutter is currently working to support sensitive context protection on Android (see https://github.com/flutter/flutter/issues/150218) which is based on the [View#setContentSensitivity](https://developer.android.com/reference/android/view/View#setContentSensitivity(int)) API that is only supported on Android 15+. We should consider implementing some form of backwards compatibility for lower API versions by trying to detect when the screen is being shared by using the [DisplayManager](https://developer.android.com/reference/android/hardware/display/DisplayManager) to detect active `Display`s.
Note that this would also require figuring out how to obscure sensitive content if we are able to detect screen sharing on Android APIs < 15.
See comment on technical limitations section in go/flutter-android-sensitive-content for more context.
|
c: new feature,platform-android,engine,P3,team-android,triaged-android
|
low
|
Minor
|
2,676,620,744 |
pytorch
|
Consider noting the PyTorch version for BC-breaking error messages
|
I hit the following error message in torch.load https://github.com/pytorch/pytorch/blob/8d708090c0eb306facfd8f85d58c578a8cbbe689/torch/serialization.py#L1338-L1341 while working on enabling a model with torch.compile for user empathy day.
It took me a couple of minutes to parse the error message and to determine if it was my fault, the code's fault, or PyTorch's fault. My conclusion was that this code used to succeed in the past but we have changed it to fail in a recent release (or on main). I would have appreciated an error message to that effect, like:
"In PyTorch 2.x we changed the behavior of torch.load to only load weights. This torch.load failed because it serialized more than just weights. If you're confident that the file is safe to load, then you may pass weights_only=False to torch.load)."
cc @ezyang @gchanan @mikaylagawarecki
|
module: bc-breaking,triaged,topic: improvements
|
low
|
Critical
|
2,676,625,500 |
tauri
|
[bug] BSOD rarely when exiting the application
|
### Describe the bug
Hello
My application causes a BSOD, error code โVideo_scheduler_internal_errorโ, very rarely (about once out of 70 times) when it is exited, either by code, by alt+F4, by reloading the code or by clicking on the exit button
The event observer also gives the indication โThe driver detected an internal driver error on \Device\VBoxNetLwf.โ
My dependencies are :
serde_json = โ1.0
serde = { version = โ1.0โ, features = [โderiveโ] }
tokio = { version = โ1โ, features = [โfullโ] }
tauri = { version = โ1.8.1โ, features = [โapi-allโ] }
reqwest = { version = โ0.11โ, features = [โjsonโ, โrustls-tlsโ] }
futures = โ0.3โ
zip = โ2.1.6โ
My application is developed on Windows 10, npm and javascript. Version of tauri-cli 2.1.0
Let me know if you need any further information
### Reproduction
Quitting the application gives it a small but real chance of causing a BSOD; it doesn't matter how you quit
### Expected behavior
Application no longer causes BSODs
### Full `tauri info` output
```text
[โ] Environment
- OS: Windows 10.0.19045 X64
โ WebView2: 130.0.2849.80
โ MSVC: Visual Studio Community 2022
โ rustc: 1.82.0 (f6e511eec 2024-10-15)
โ cargo: 1.82.0 (8f40fc59f 2024-08-21)
โ rustup: 1.27.1 (54dd3d00f 2024-04-24)
โ Rust toolchain: stable-x86_64-pc-windows-msvc (default)
- node: 21.7.1
- npm: 9.1.2
[-] Packages
- tauri [RUST]: 1.8.1
- tauri-build [RUST]: 1.5.5
- wry [RUST]: 0.24.11
- tao [RUST]: 0.16.10
- @tauri-apps/api [NPM]: 1.6.0 (outdated, latest: 2.1.1)
- @tauri-apps/cli [NPM]: 1.6.3 (outdated, latest: 2.1.0)
[-] App
- build-type: bundle
- CSP: unset
- distDir: ../dist
- devPath: http://localhost:1420/
- framework: Vue.js
- bundler: Vite
```
### Stack trace
_No response_
### Additional context
_No response_
|
type: bug,status: needs triage
|
low
|
Critical
|
2,676,713,163 |
angular
|
Select tag is flickering SSR
|
### Which @angular/* package(s) are the source of the bug?
Don't known / other
### Is this a regression?
Yes
### Description
As you can see on the video, the select tag is flickering compared to input tag, I use SSR
### Please provide a link to a minimal reproduction of the bug
https://stackblitz.com/~/github.com/illunix/flickering-issue-angular
### Please provide the exception or error you saw
https://github.com/user-attachments/assets/71172d3c-bb6b-462b-8738-84b3aa88419e
### Please provide the environment you discovered this bug in (run `ng version`)
```true
Angular CLI: 19.0.0
Node: 20.16.0
Package Manager: npm 10.9.0
OS: win32 x64
```
### Anything else?
_No response_
|
workaround2: non-obvious,area: forms,area: server,P3
|
low
|
Critical
|
2,676,713,418 |
go
|
build: build failure on gotip-linux-loong64
|
```
#!watchflakes
default <- builder == "gotip-linux-loong64" && repo == "go" && mode == "build"
```
Issue created automatically to collect these failures.
Example ([log](https://ci.chromium.org/b/8730734133647122513)):
[W2024-11-21T00:44:36.689641+08:00 1080691 0 json_subtract.go:121] Unknown fields while parsing property namespace "": {"is_google":false,"mode":0}
2024/11/21 00:44:36 run starting
2024/11/21 00:44:37 installed tools
โ [watchflakes](https://go.dev/wiki/Watchflakes)
|
NeedsInvestigation
|
low
|
Critical
|
2,676,749,008 |
vscode
|
Quick search does not retain search term when moving to view
|
Notice the term isn't carried over to the view:

|
bug,search
|
low
|
Minor
|
2,676,749,907 |
pytorch
|
Spsolve for sparse, complex matrices
|
### ๐ Describe the bug
Sparse linear algebra matrix solver doesn't work with complex numbers on a GPU (solver not available at all on the CPU)
Sample Code:
```
import torch
from torch import sparse
n = 10
A = torch.rand(n, n).type(torch.complex128).to_sparse_csr().cuda()
b = torch.rand(n).type(torch.complex128).cuda()
return sparse.spsolve(A, b)
```
Traceback:
```
File "test.py", line 36, in solve_matrix_equation
return sparse.spsolve(A, b)
RuntimeError: "create_matrix" not implemented for 'ComplexDouble'
```
### Versions
Collecting environment information...
PyTorch version: 2.6.0a0+git99a0321
Is debug build: False
CUDA used to build PyTorch: 12.4
ROCM used to build PyTorch: N/A
OS: Rocky Linux release 8.10 (Green Obsidian) (x86_64)
GCC version: (GCC) 8.5.0 20210514 (Red Hat 8.5.0-22)
Clang version: Could not collect
CMake version: version 3.31.0
Libc version: glibc-2.28
Python version: 3.10.15 (main, Oct 3 2024, 07:27:34) [GCC 11.2.0] (64-bit runtime)
Python platform: Linux-4.18.0-553.8.1.el8_10.x86_64-x86_64-with-glibc2.28
Is CUDA available: True
CUDA runtime version: 12.4.131
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: Tesla V100-SXM2-16GB
Nvidia driver version: 555.42.06
cuDNN version: Probably one of the following:
/usr/lib64/libcudnn.so.8.9.7
/usr/lib64/libcudnn_adv_infer.so.8.9.7
/usr/lib64/libcudnn_adv_train.so.8.9.7
/usr/lib64/libcudnn_cnn_infer.so.8.9.7
/usr/lib64/libcudnn_cnn_train.so.8.9.7
/usr/lib64/libcudnn_ops_infer.so.8.9.7
/usr/lib64/libcudnn_ops_train.so.8.9.7
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 8
On-line CPU(s) list: 0-7
Thread(s) per core: 2
Core(s) per socket: 4
Socket(s): 1
NUMA node(s): 1
Vendor ID: GenuineIntel
CPU family: 6
Model: 79
Model name: Intel(R) Xeon(R) CPU E5-2686 v4 @ 2.30GHz
Stepping: 1
CPU MHz: 2703.014
CPU max MHz: 3000.0000
CPU min MHz: 1200.0000
BogoMIPS: 4600.03
Hypervisor vendor: Xen
Virtualization type: full
L1d cache: 32K
L1i cache: 32K
L2 cache: 256K
L3 cache: 46080K
NUMA node0 CPU(s): 0-7
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single pti fsgsbase bmi1 hle avx2 smep bmi2 erms invpcid rtm rdseed adx xsaveopt
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.26.4
[pip3] optree==0.13.1
[pip3] torch==2.6.0a0+git99a0321
[conda] blas 1.0 mkl
[conda] magma-cuda124 2.6.1 1 pytorch
[conda] mkl 2023.1.0 h213fc3f_46344
[conda] mkl-include 2025.0.1 pypi_0 pypi
[conda] mkl-service 2.4.0 py310h5eee18b_1
[conda] mkl-static 2025.0.1 pypi_0 pypi
[conda] mkl_fft 1.3.11 py310h5eee18b_0
[conda] mkl_random 1.2.8 py310h1128e8f_0
[conda] numpy 1.26.4 py310h5f9d8c6_0
[conda] numpy-base 1.26.4 py310hb5e798b_0
[conda] optree 0.13.1 pypi_0 pypi
[conda] torch 2.6.0a0+git99a0321 pypi_0 pypi
cc @alexsamardzic @nikitaved @pearu @cpuhrsch @amjames @bhosmer @jcaip
|
module: sparse,triaged
|
low
|
Critical
|
2,676,763,197 |
vscode
|
Provide quick fix for missing copyright statement in CSS files
|
TS file:

CSS file:

|
feature-request,engineering
|
low
|
Minor
|
2,676,770,966 |
next.js
|
All imports must resolve in `next.config` when running `next lint`. This prevents monorepos from linting before building dependent packages.
|
### Link to the code that reproduces this issue
https://codesandbox.io/p/devbox/damp-sea-q85m6v
### To Reproduce
`npm run lint`
### Current vs. Expected behavior
`Error [ERR_PACKAGE_PATH_NOT_EXPORTED]: No "exports" main defined in /project/sandbox/node_modules/some-package/package.json imported from /project/sandbox/next.config.mjs`
It should be possible to lint a project _without_ a package being used as a plugin in the next config being built. In monorepo pipelines it's common to lint prior to building all dependent packages.
Running `eslint` independently in a monorepo does not complain when a package referenced by a `file:` resolution in `package.json` points to non-existent files; this is a run-time not a lint-time concern.
### Provide environment information
```bash
Operating System:
Platform: linux
Arch: x64
Version: #1 SMP PREEMPT_DYNAMIC Sun Aug 6 20:05:33 UTC 2023
Available memory (MB): 4102
Available CPU cores: 2
Binaries:
Node: 20.9.0
npm: 9.8.1
Yarn: 1.22.19
pnpm: 8.10.2
Relevant Packages:
next: 15.0.4-canary.20 // Latest available version is detected (15.0.4-canary.20).
eslint-config-next: N/A
react: 19.0.0-rc-380f5d67-20241113
react-dom: 19.0.0-rc-380f5d67-20241113
typescript: 5.3.3
Next.js Config:
output: N/A
```
### Which area(s) are affected? (Select all that apply)
Developer Experience, Linting
### Which stage(s) are affected? (Select all that apply)
next dev (local)
### Additional context
_No response_
|
bug,Linting
|
low
|
Critical
|
2,676,780,132 |
flutter
|
Deprecate and remove Dart-side `DeferredComponents`
|
Flutter supports two sets of features around deferring code: (a) deferring chunks of Dart and (b) deferring the engine. The latter (b) is used, including in google3, to, quote _"not pay 5mb when a 1000 screen app adds their first 1 Flutter screen"_. The former feature, however, is not known to have any internal use, and we suspect very little (or no) external use.
Feature overview:
- https://github.com/flutter/flutter/wiki/Deferred-Components
- https://docs.flutter.dev/perf/deferred-components
From talking to @jonahwilliams and in my own experiences spelunking through the `flutter` tool, deferred components is a complex feature that is almost not tested at all, and has implementation elements throughout most of the product - in the tool, framework, and engine. Deprecating and removing it would remove a significant source of [haunted graveyard](https://www.usenix.org/sites/default/files/conference/protected-files/srecon17americas_slides_reese.pdf)-style complexity, especially if the feature is not used or tested.
As a first step, I'll work with @andrewkolos to enable Google analytics tracking for this feature.
|
team,platform-android,c: API break,c: proposal,P2,c: tech-debt,fyi-android,fyi-tool
|
low
|
Minor
|
2,676,789,775 |
PowerToys
|
Screen Ruler Crashing
|
### Microsoft PowerToys version
0.86.0
### Installation method
PowerToys auto-update
### Running as admin
None
### Area(s) with issue?
Screen ruler
### Steps to reproduce
STEP 1:
- Go down to the Windows10 taskbar and click on the PowerToys icon.
- Click on Screen Ruler from the quick access menu.
OR
- Use the default Win + Shift + m
STEPs 2-4:
- This itself doesn't always bring up the tool, but when it does press the Vertical Spacing button.
- If it hasn't frozen yet, click anywhere on your screen.
- Without fail it will freeze and become unresponsive requiring killing the process,
### โ๏ธ Expected Behavior
For the tool to open every time I press the icon from the taskbar or use the keyboard shortcut and for the tool not to crash.
### โ Actual Behavior
The tool inconsistently opens using any regular method of opening it and always crashes when it does manage to open within a very short amount of time (seemingly after one action and then clicking off).
### Other Software
[PowerToysReport_2024-11-20-12-25-12.zip](https://github.com/user-attachments/files/17835262/PowerToysReport_2024-11-20-12-25-12.zip)
|
Issue-Bug,Needs-Triage
|
low
|
Critical
|
2,676,790,575 |
kubernetes
|
fix device manager permafailing test
|
### Which jobs are failing?
https://testgrid.k8s.io/sig-node-presubmits#pr-node-kubelet-serial-containerd-flaky
### Which tests are failing?
E2eNode Suite: [It] [sig-node] Device Plugin [NodeFeature:DevicePlugin] [Serial] DevicePlugin [Serial] [Disruptive] Keeps device plugin assignments across node reboots (no pod restart, no device plugin re-registration) [Flaky]
### Since when has it been failing?
Oct 30 2024, likely before https://github.com/kubernetes/kubernetes/issues/128114
### Testgrid link
https://prow.k8s.io/view/gs/kubernetes-ci-logs/pr-logs/directory/pull-kubernetes-node-kubelet-containerd-flaky/1851562955540271104
### Reason for failure (if possible)
need to review the test purpose and update or remove it, with a rationale
### Anything else we need to know?
followup of https://github.com/kubernetes/kubernetes/issues/128443
put issue on me to make sure we followup and we review the tests
### Relevant SIG(s)
/sig
|
priority/important-soon,sig/node,kind/failing-test,triage/accepted
|
low
|
Critical
|
2,676,812,466 |
next.js
|
Inconsistent behaviour of revalidatePath between local and Vercel
|
### Link to the code that reproduces this issue
https://github.com/damianfrizzi/next-revalidate-bug
### To Reproduce
- Go to [https://next-revalidate-bug.vercel.app](https://t.co/BBukGxf9iY) and note the number at the top
- Navigate to "Go to some page" and note the number at the top
- Go back to homepage and click "Invalidate German" and refresh page
- Number correctly changes on the homepage
- BUT number also changes on the "/somePage" route. This shouldn't happen
### Current vs. Expected behavior
`revalidatePath` behaves correctly when using `bun build && bun start` on my local machine. I.e., invalidating the homepage doesn't invalidate any other (sub) pages.
When deployed to Vercel, invalidating the homepage also invalidates other (sub) pages.
### Provide environment information
```bash
Operating System:
Platform: darwin
Arch: arm64
Version: Darwin Kernel Version 24.0.0: Tue Sep 24 23:39:07 PDT 2024; root:xnu-11215.1.12~1/RELEASE_ARM64_T6000
Available memory (MB): 32768
Available CPU cores: 10
Binaries:
Node: 20.17.0
npm: 10.8.2
Yarn: 1.22.19
pnpm: N/A
Relevant Packages:
next: 15.0.3 // Latest available version is detected (15.0.3).
eslint-config-next: 15.0.3
react: 19.0.0-rc-66855b96-20241106
react-dom: 19.0.0-rc-66855b96-20241106
typescript: 5.6.3
Next.js Config:
output: N/A
```
### Which area(s) are affected? (Select all that apply)
Internationalization (i18n), Middleware, Runtime
### Which stage(s) are affected? (Select all that apply)
Vercel (Deployed)
### Additional context
I'm using `trailingSlash: true` in my next.config. I suspect that there is an error in connection with this configuration.
|
bug,Middleware,Internationalization (i18n),Runtime
|
low
|
Critical
|
2,676,884,649 |
go
|
crypto/ecdsa: Sign() panics if public key is not set
|
### Go version
go version devel go1.24-7c7170e Wed Nov 20 18:27:31 2024 +0000 linux/amd64
### Output of `go env` in your module/workspace:
```shell
AR='ar'
CC='clang-15'
CGO_CFLAGS='-O2 -g'
CGO_CPPFLAGS=''
CGO_CXXFLAGS='-O2 -g'
CGO_ENABLED='1'
CGO_FFLAGS='-O2 -g'
CGO_LDFLAGS='-O2 -g'
CXX='clang++-15'
GCCGO='gccgo'
GO111MODULE=''
GOAMD64='v1'
GOARCH='amd64'
GOAUTH='netrc'
GOBIN=''
GOCACHE='/home/jhg/.cache/go-build'
GODEBUG=''
GOENV='/home/jhg/.config/go/env'
GOEXE=''
GOEXPERIMENT=''
GOFIPS140='off'
GOFLAGS=''
GOGCCFLAGS='-fPIC -m64 -pthread -fno-caret-diagnostics -Qunused-arguments -Wl,--no-gc-sections -fmessage-length=0 -ffile-prefix-map=/tmp/go-build1928511716=/tmp/go-build -gno-record-gcc-switches'
GOHOSTARCH='amd64'
GOHOSTOS='linux'
GOINSECURE=''
GOMOD='/home/jhg/oss-fuzz-379773077/p/go.mod'
GOMODCACHE='/home/jhg/oss-fuzz-379773077/go-dev/packages/pkg/mod'
GONOPROXY=''
GONOSUMDB=''
GOOS='linux'
GOPATH='/home/jhg/oss-fuzz-379773077/go-dev/packages'
GOPRIVATE=''
GOPROXY='https://proxy.golang.org,direct'
GOROOT='/home/jhg/oss-fuzz-379773077/go-dev'
GOSUMDB='sum.golang.org'
GOTELEMETRY='local'
GOTELEMETRYDIR='/home/jhg/.config/go/telemetry'
GOTMPDIR=''
GOTOOLCHAIN='auto'
GOTOOLDIR='/home/jhg/oss-fuzz-379773077/go-dev/pkg/tool/linux_amd64'
GOVCS=''
GOVERSION='devel go1.24-7c7170e Wed Nov 20 18:27:31 2024 +0000'
GOWORK=''
PKG_CONFIG='pkg-config'
```
### What did you do?
```go
package main
import (
"crypto/ecdsa"
"crypto/elliptic"
"crypto/rand"
"crypto/sha256"
"fmt"
"math/big"
)
func main() {
priv, ok := new(big.Int).SetString("123", 10)
if ok == false {
panic("Cannot decode bignum")
}
var priv_ecdsa ecdsa.PrivateKey
priv_ecdsa.D = priv
priv_ecdsa.PublicKey.Curve = elliptic.P256()
msg := "hello, world"
hash := sha256.Sum256([]byte(msg))
r, s, err := ecdsa.Sign(rand.Reader, &priv_ecdsa, hash[:])
fmt.Println(err)
fmt.Println(r.String())
fmt.Println(s.String())
}
```
### What did you see happen?
```
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x10 pc=0x4c2117]
goroutine 1 [running]:
math/big.(*Int).Sign(...)
/home/jhg/oss-fuzz-379773077/go-dev/src/math/big/int.go:48
crypto/ecdsa.pointFromAffine({0x56bc88?, 0x646070?}, 0x0, 0x0)
/home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:402 +0x37
crypto/ecdsa.privateKeyToFIPS[...](0xc0002d02c0, 0xc0002dff18)
/home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:391 +0x38
crypto/ecdsa.signFIPS[...](0xc0002d02c0, 0xc0002c8df8, {0x56b020?, 0x66c280}, {0xc0002dfed8, 0x20, 0x20})
/home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:234 +0x46
crypto/ecdsa.SignASN1({0x56b020, 0x66c280}, 0xc0002dff18, {0xc0002dfed8, 0x20, 0x20})
/home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa.go:220 +0x229
crypto/ecdsa.Sign({0x56b020?, 0x66c280?}, 0xbf7cbadf074d6483?, {0xc0002c8ed8?, 0xc0002c8ef8?, 0xc?})
/home/jhg/oss-fuzz-379773077/go-dev/src/crypto/ecdsa/ecdsa_legacy.go:60 +0x37
main.main()
/home/jhg/oss-fuzz-379773077/p/x.go:23 +0xf4
```
### What did you expect to see?
No panic and output like:
```
<nil>
40753909606936490861524166827361514810969838335424734688996005945565630866707
3003946056421339230760555414094068582110502790139132375823642300394845145720
```
|
NeedsInvestigation
|
low
|
Critical
|
2,676,906,576 |
fastapi
|
Will FastAPI support QUERY http method? "app.query"
|
### Discussed in https://github.com/fastapi/fastapi/discussions/6049
<div type='discussions-op-text'>
<sup>Originally posted by **FilipeMarch** December 16, 2022</sup>
### First Check
- [X] I added a very descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the FastAPI documentation, with the integrated search.
- [X] I already searched in Google "How to X in FastAPI" and didn't find any information.
- [X] I already read and followed all the tutorial in the docs and didn't find an answer.
- [X] I already checked if it is not related to FastAPI but to [Pydantic](https://github.com/samuelcolvin/pydantic).
- [X] I already checked if it is not related to FastAPI but to [Swagger UI](https://github.com/swagger-api/swagger-ui).
- [X] I already checked if it is not related to FastAPI but to [ReDoc](https://github.com/Redocly/redoc).
### Commit to Help
- [X] I commit to help with one of those options ๐
### Example Code
```python
from fastapi import FastAPI
app = FastAPI()
@app.query('/query/subjects')
def query_subjects(schema: ArbitrarySchema):
with Session(engine) as db:
subjects = db.query(Subject).all()
return schema(**subjects)
# something like this
```
### Description
There is a new HTTP method called QUERY, I discovered it this week and it is super interesting!
https://www.ietf.org/archive/id/draft-ietf-httpbis-safe-method-w-body-02.html
I was trying to make a GET route that would accept a request body, but started receiving an error on Swagger UI:
[TypeError: Request has method 'GET' and cannot have a body](https://github.com/swagger-api/swagger-ui/issues/5891#issuecomment-1340591857)
The idea is that sometimes we need to make a big or complex query, and this is the scenario in which we can see the advantages of GraphQL, and although I have seen that [we can integrate FastAPI with GraphQL](https://fastapi.tiangolo.com/advanced/graphql/), I was wondering if FastAPI will ever be able to simply accept a QUERY method like, for example, `app.query("/query/subjects")`.
For example, suppose I have this object:
```yml
{
"id": 1,
"name": "Math"
"tags": [
{
"id": 1,
"name": "Algebra",
"number_of_clicks": 1,
"number_of_questions": 7,
"number_of_answers": 3,
"number_of_comments": 2,
"number_of_votes": 1,
}]
"topics": [
{
"id": 1,
"name": "Linear Equations",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_tutorials": 1,
"number_of_questions": 7,
"posts": [
{
"id": 1,
"title": "How to solve linear equations?",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_answers": 3,
"number_of_comments": 2,
"number_of_votes": 1,
"answers": [
{
"id": 1,
"content": "You can solve linear equations by using the substitution method.",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_comments": 2,
"number_of_votes": 1,
"comments": [
{
"id": 1,
"content": "That's a great answer!",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_votes": 1,
}"
]
}
]
}
]
}
]
}
```
I want the client to be able to QUERY this in any specific way, maybe he wants GET objects with this format
```yml
{
"id": 1,
"name": "Math"
"tags": [
{
"id": 1,
"name": "Algebra",
}]
"topics": [
{
"id": 1,
"name": "Linear Equations",
}
]
}
```
But it could be totally different, and I can't predict the way the client is gonna query it and make a new route for every combination and possible schema for a complex object. The QUERY route would accept a request body, and then we would filter the response with the schema the client sends on request body.
- Would it be too complicated to implement this on FastAPI?
- Is there any hope that this will someday be implemented in FastAPI? XD
Thanks!
### Wanted Solution
...
### Wanted Code
```python
...
```
### Alternatives
...
### Operating System
macOS
### Operating System Details
_No response_
### FastAPI Version
A
### Python Version
3.11
### Additional Context
_No response_</div>
|
feature,good first issue
|
low
|
Critical
|
2,676,962,886 |
vscode
|
Feature Request: Expose findWidget Actions State via API
|
The VS Code Find Widget allows toggling control options like `toggleFindCaseSensitive`, `toggleFindWholeWord`, `toggleFindRegex`. However, there is no way to programmatically query the current state of these options via the extension API (that I'm aware of). This limitation prevents extensions from seamlessly integrating with or enhancing the Find Widgetโs functionality.
### Proposed Solution
Expose the current state of the Find Widget toggles (e.g., `findCaseSensitive`, `findWholeWord`, `findRegex`) through the extension API.
These could be added to the `activeTextEditor` options either nested or directly:
`vscode.window.activeTextEditor.options.findWidget.findCaseSensitive`
An alternative would be to allow toggling to an explicit state:
`vscode.commands.executeCommand("toggleFindCaseSensitive", true)`
### Workarounds / Context
I've thus far not been able to find a way to work around this- I am attempting to write an extension that modifies the cmd+d selection functionality to more closely match Atom/Pulsar in such a way that it determines if you are trying to select a full word (cursor on the word, nothing actively selected) vs partial matches (highlighted/selected text first, THEN cmd+d) but because there is no way to check these settings I am unable to toggle them to the desired state based on context.
|
feature-request,api
|
low
|
Minor
|
2,676,968,806 |
deno
|
Upgrade to TypeScript 5.7
|
We're going to push this back a release to Deno 2.2 due to https://devblogs.microsoft.com/typescript/announcing-typescript-5-7-beta/#typedarrays-are-now-generic-over-arraybufferlike in order to find a better migration strategy.
|
feat,tsc
|
low
|
Minor
|
2,677,039,656 |
ui
|
[bug]: Tooltip on Themes page doesn't inherit theme
|
### Describe the bug
On the Themes page:
https://ui.shadcn.com/themes
When hovering over the new message button in the ChatDemo, the Tooltip portals in on the body, which is outside the .theme-red selector so the tooltip doesn't inherit the theme.
I think one potential solution could be to change the tooltip.tsx component to portal the tooltip content in relative to the tooltip root vs. on document.body.
### Affected component/components
Documentation, Tooltip
### How to reproduce
1. Go to https://ui.shadcn.com/themes
2. Hover on new message button in ChatDemo
3. Notice tooltip is black, when it should be red if the red theme is selected.
### Codesandbox/StackBlitz link
https://ui.shadcn.com/themes
### Logs
_No response_
### System Info
```bash
Mac, Chrome
```
### Before submitting
- [X] I've made research efforts and searched the documentation
- [X] I've searched for existing issues
|
bug
|
low
|
Critical
|
2,677,041,726 |
PowerToys
|
Cannot launch UI elements: faulting module Microsoft.UI.Xaml.dll
|
### Microsoft PowerToys version
0.86.0
### Installation method
GitHub
### Running as admin
No
### Area(s) with issue?
General, Settings, Advanced Paste, PowerToys Run
### Steps to reproduce
Installed 0.86.0 (machine-wide) from GitHub on W11 23H2 22631 (multiple machines), both manually and through SCCM.
I get the following error when trying to launch any UI part of PowerToys:
```
Faulting application name: PowerToys.PowerLauncher.exe, version: 0.86.0.0, time stamp: 0x66e80000
Faulting module name: KERNELBASE.dll, version: 10.0.22621.4391, time stamp: 0x7433a115
Exception code: 0xe0434352
Fault offset: 0x000000000005fa4c
Faulting process ID: 0x0xbb0
Faulting application start time: 0x0x1db3b83c903a38c
Faulting application path: C:\Program Files\PowerToys\PowerToys.PowerLauncher.exe
Faulting module path: C:\WINDOWS\System32\KERNELBASE.dll
```
```
Faulting application name: PowerToys.Settings.exe, version: 0.86.0.0, time stamp: 0x66e80000
Faulting module name: Microsoft.UI.Xaml.dll, version: 3.1.6.0, time stamp: 0x0caad657
Exception code: 0xc000027b
Fault offset: 0x0000000000009125
Faulting process ID: 0x0xf24
Faulting application start time: 0x0x1db3b83caefa763
Faulting application path: C:\Program Files\PowerToys\WinUI3Apps\PowerToys.Settings.exe
Faulting module path: C:\Program Files\PowerToys\WinUI3Apps\Microsoft.UI.Xaml.dll
```
```
Faulting application name: PowerToys.Peek.UI.exe, version: 0.86.0.0, time stamp: 0x66e80000
Faulting module name: Microsoft.UI.Xaml.dll, version: 3.1.6.0, time stamp: 0x0caad657
Exception code: 0xc000027b
Fault offset: 0x0000000000009125
Faulting process ID: 0x0x1b20
Faulting application start time: 0x0x1db3b84ec20d9a2
Faulting application path: C:\Program Files\PowerToys\WinUI3Apps\PowerToys.Peek.UI.exe
Faulting module path: C:\Program Files\PowerToys\WinUI3Apps\Microsoft.UI.Xaml.dll
```
---
Trying to generate a debug report results in (likely because diagnostics is disabled through GPO):
```
Application: PowerToys.PowerLauncher.exe
CoreCLR Version: 8.0.1024.46610
.NET Version: 8.0.10
Description: The process was terminated due to an unhandled exception.
Exception Info: System.InvalidCastException: Unable to cast object of type 'System.Byte[]' to type 'System.Int32'.
at Microsoft.PowerToys.Telemetry.DataDiagnosticsSettings.GetEnabledValue()
at Microsoft.PowerToys.Telemetry.ETWTrace..ctor()
at PowerLauncher.App..ctor()
at PowerLauncher.App.Main()
```
### โ๏ธ Expected Behavior
Any UI element of PowerToys opens as expected.
### โ Actual Behavior
Nothing opens (the icon is visible in the system tray), and occasionally, the system hangs (especially when trying to do something with advanced paste).
Interestingly, Always On Top works perfectly. Running the application as an administrator also works perfectly (no GPO enforced).
### Other Software
_No response_
|
Issue-Bug,Resolution-Fix Committed
|
medium
|
Critical
|
2,677,049,900 |
react
|
[React 19] Regression when using `createPortal` with DOM element created by `dangerouslySetInnerHTML`
|
In React 18, it was possible to use `createPortal` with a DOM element created by `dangerouslySetInnerHTML`.
Example (adapted from [this Stack Overflow answer](https://stackoverflow.com/questions/62300569/add-react-component-in-html-set-by-dangerouslysetinnerhtml/77099814#77099814)):
```tsx
import { useCallback, useState } from "react";
import { createPortal } from "react-dom";
export default function App() {
const htmlFromElsewhere = `foo <span class="portal-container"></span> bar`;
return <InnerHtmlWithPortals html={htmlFromElsewhere} />
}
function InnerHtmlWithPortals({ html }: { html: string }) {
const [portalContainer, setPortalContainer] = useState<Element | null>(null)
const refCallback = useCallback((el: HTMLDivElement) => {
setPortalContainer(el?.querySelector(".portal-container"))
})
return <>
<div ref={refCallback} dangerouslySetInnerHTML={{ __html: html }} />
{portalContainer && createPortal(<Cake />, portalContainer)}
</>
}
function Cake() {
return <strong>cake</strong>
}
```
React 18 CodeSandbox: https://codesandbox.io/p/sandbox/optimistic-kowalevski-73sk5w
In React 19, this no longer works. React appears to be re-rendering the inner HTML after calling `refCallback`. Thus, `createPortal` succeeds, but the `portalContainer` element that it uses is no longer part of the DOM.
React 19 CodeSandbox: https://codesandbox.io/p/sandbox/vibrant-cloud-gd8yzr
It is possible to work around the issue by setting `innerHTML` directly instead of using `dangerouslySetInnerHTML`:
```diff
const [portalContainer, setPortalContainer] = useState<Element | null>(null)
const refCallback = useCallback((el: HTMLDivElement) => {
+ if (el) el.innerHTML = html
setPortalContainer(el?.querySelector(".portal-container"))
- })
+ }, [html])
return <>
- <div ref={refCallback} dangerouslySetInnerHTML={{ __html: html }} />
+ <div ref={refCallback} />
{portalContainer && createPortal(<Cake />, portalContainer)}
</>
```
But I'm not sure whether that is a reliable solution.
|
React 19
|
medium
|
Minor
|
2,677,073,607 |
kubernetes
|
[FG:InPlacePodVerticalScaling] pull-kubernetes-e2e-capz-windows-master test fail with InPlacePodVerticalScaling Beta
|
### Which jobs are failing?
pull-kubernetes-e2e-capz-windows-master
https://prow.k8s.io/view/gs/kubernetes-ci-logs/pr-logs/pull/128880/pull-kubernetes-e2e-capz-windows-master/1859240005449289728
### Which tests are failing?
Kubernetes e2e suite: [It] [sig-storage]
- Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]
-Secrets should be consumable from pods in volume [NodeConformance] [Conformance]
-ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 23s
-Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 33s
-Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 25s
- Projected secret should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] expand_more 25s
- Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] expand_more 27s
- Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] expand_more 31s
- Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] expand_more 27s
- Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] expand_more 27s
- Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] expand_more 25s
- Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] expand_more 27s
- ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] expand_more 1m9s
- ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] expand_more 25s
- ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] expand_more 27s
- Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] expand_more
Kubernetes e2e suite: [It] [sig-windows] [Feature:Windows]
- Windows volume mounts check volume mount permissions container should have readOnly permissions on emptyDir expand_more 35s
Kubernetes e2e suite: [It] [sig-node]
- Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] expand_more 10m3s
### Since when has it been failing?
Last successful 12 November 20:20 1856416666905219072
https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-capz-master-windows/1856416666905219072
First failed 12 November 23:20 1856461965182898176
https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-kubernetes-e2e-capz-master-windows/1856461965182898176
### Testgrid link
https://testgrid.k8s.io/sig-release-master-informing#capz-windows-master
### Reason for failure (if possible)
_No response_
### Anything else we need to know?
_No response_
### Relevant SIG(s)
/sig-node /sig-storage /sig-windows
|
priority/important-soon,sig/node,sig/windows,kind/failing-test,triage/accepted
|
medium
|
Critical
|
2,677,080,060 |
go
|
proposal: hash: standardize the hash function
|
**Background**: Issue https://github.com/golang/go/issues/69420 proposes a `types.Hash` function that provides a hash operator for `types.Type` values that is consistent with the equivalence relation for types defined by `types.Identical`. The only real question in that proposal is: what is to be Go's future convention for the signature of a custom hash function?
The naive answer is `func(T) uint64`, where T stands for `types.Type`. However, hash tables with only a single hash function are vulnerable to a "hash flooding" denial-of-service attack, in which an adversary provides many inputs that all hash to the same value, causing the table to degenerate into a linked list. The defense is for each process (or particular hash table) to select a different hash function at random so that the values cannot be predicted. The `hash/maphash` package embodies this approach: a random Seed value determines the hash functions (`Hash.Strings` and `Hash.Bytes`), which depend on the seed's opaque random value. (maphash does not provide a Seed-dependent way to hash integers.)
So, perhaps custom hash functions should accept a `maphash.Seed` parameter. And perhaps Seed should also provide methods for hashing integers.
**Proposal:** In the hash package, define function types HashFunc and EqFunc that document the conventions for Go hash tables.
Here is the minimal answer, without Seeds:
```go
package hash
// HashFunc defines a hash function for values of type T.
// A conformant HashFunc and EqFunc pair (hash, eq) must observe the invariant
// that eq(x, y) => hash(x) == hash(y).
type HashFunc[T any] func(T) uint64
// EqFunc defines an equivalence relation for values of type T.
type EqFunc[T any] func(x, y T) bool
```
@ianlancetaylor @griesemer
|
Proposal
|
medium
|
Major
|
2,677,083,268 |
pytorch
|
Dynamo should reorder warnings (and other logging) by default if they don't involve tensor values
|
The thing I ran into is https://gist.github.com/zou3519/4c45db951c17845a598f4e9d9f433ba9 (instancenorm warns if an input looks wrong). One can also run into this via deprecation warnings.
cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @chauhang @amjames
|
triaged,oncall: pt2,module: dynamo,empathy-day
|
low
|
Minor
|
2,677,127,597 |
rust
|
Tracking Issue for OsString::truncate
|
<!--
Thank you for creating a tracking issue!
Tracking issues are for tracking a feature from implementation to stabilization.
Make sure to include the relevant RFC for the feature if it has one.
If the new feature is small, it may be fine to skip the RFC process. In that
case, you can use `issue = "none"` in your initial implementation PR. The
reviewer will ask you to open a tracking issue if they agree your feature can be
added without an RFC.
-->
Feature gate: `#![feature(os_string_truncate)]`
This is a tracking issue for the
<!--
Include a short description of the feature.
-->
### Public API
<!--
For most library features, it'd be useful to include a summarized version of the public API.
(E.g. just the public function signatures without their doc comments or implementation.)
-->
```rust
// std::ffi::os_str
impl OsString {
pub fn truncate(&mut self, new_len: usize);
}
```
### Steps / History
<!--
For larger features, more steps might be involved.
If the feature is changed later, please add those PRs here as well.
-->
- [x] ACP Approved: https://github.com/rust-lang/libs-team/issues/486
- [ ] Implementation:
- [ ] Final comment period (FCP)[^1]
- [ ] Stabilization PR
<!--
Once the feature has gone through a few release cycles and there are no
unresolved questions left, the feature might be ready for stabilization.
If this feature didn't go through the RFC process, a final comment period
(FCP) is always needed before stabilization. This works as follows:
A library API team member can kick off the stabilization process, at which point
the rfcbot will ask all the team members to verify they agree with
stabilization. Once enough members agree and there are no concerns, the final
comment period begins: this issue will be marked as such and will be listed
in the next This Week in Rust newsletter. If no blocking concerns are raised in
that period of 10 days, a stabilization PR can be opened by anyone.
-->
### Unresolved Questions
<!--
Include any open questions that need to be answered before the feature can be
stabilised. If multiple (unrelated) big questions come up, it can be a good idea
to open a separate issue for each, to make it easier to keep track of the
discussions.
It's useful to link any relevant discussions and conclusions (whether on GitHub,
Zulip, or the internals forum) here.
-->
- None yet.
[^1]: https://std-dev-guide.rust-lang.org/feature-lifecycle/stabilization.html
|
T-libs-api,C-tracking-issue
|
low
|
Minor
|
2,677,131,714 |
flutter
|
Revisit inline vs out-of-line implementations of methods in Impeller geometry classes
|
Currently many of the Impeller geometry classes are templates and so they must implement all of their methods inline, but some of them (for instance `Matrix`) are regular classes and so the methods could be implemented in their associated `.cc` file. We've been shifting the code to more cases of moving implementations to the `.cc` file and so we should revisit the various geometry classes to determine what's best for them.
|
engine,P3,e: impeller,team-engine,triaged-engine
|
low
|
Minor
|
2,677,146,162 |
next.js
|
URL's search params get encoded on page load.
|
### Link to the code that reproduces this issue
https://github.com/wkd-kapsule/next-router-demo
### To Reproduce
1. Go to the [production site.](https://next-router-demo-eight.vercel.app/)
2. Use a special character in the URL's search params, such as `/`. Whether by directly typing it or clicking the button on the page.
3. Watch the URL being encoded before your eyes as the page loads.
### Current vs. Expected behavior
### **_Current behavior:_**
Directly typing a url or using `useRouter` will encode the URL's search params after page loading.
### **_Expected behavior:_**
I'd like to be able to use characters such as `/` or `$` in a URL's search params.
### **_Context:_**
I use URL seach params to store some state in my app. The thing is that it's a path, thus containing `/`. In dev mode, everything is fine but somehow in production, `/` gets encoded into `%2F` once the page has loaded. It does it whether I directly type the url and press Enter or click on button with `useRouter().push`.
I've reproduced this in a minimal repo but for some reason it works fine when using `useRouter().push`. The URL gets encoded only when you manually type it and press Enter. There is basically no difference in the way I use `useRouter().push()` in this repo and in my app, I can't grasp why both behaves differently.
**NOTE: I know that I can just decode it to get the original character but my issue is more visual than technical. Having encoded characters could discourage the users of my app to use the URL to navigate. Whereas this is exactly why I put that path in the URL, so the user can directly go where they need, without having to use the interface.**
Here's the [demo](https://next-router-demo-eight.vercel.app/)
https://github.com/user-attachments/assets/4d63d9d5-f00f-4c84-b6ef-19600627b2b3
### Provide environment information
```bash
Operating System:
Platform: win32
Arch: x64
Version: Windows 11 Home
Available memory (MB): 16272
Available CPU cores: 8
Binaries:
Node: 20.13.1
npm: 10.8.1
Yarn: N/A
pnpm: 8.15.5
Relevant Packages:
next: 15.0.4-canary.19 // There is a newer canary version (15.0.4-canary.20) available, please upgrade!
eslint-config-next: 15.0.3
react: 19.0.0-rc-69d4b800-20241021
react-dom: 19.0.0-rc-69d4b800-20241021
typescript: 5.6.2
Next.js Config:
output: N/A
```
### Which area(s) are affected? (Select all that apply)
Navigation
### Which stage(s) are affected? (Select all that apply)
Vercel (Deployed)
### Additional context
I couldn't pin which Next.js versions first caused the issue because it actually occured a while ago. I didn't report it at the time as being constantly on the latest canary version, I just imagined it would eventually get solved. Now that's been several months already so I thought I better report it.
Nonetheless, I can tell you that I'm using Google Chrome and that everything works as expected in Microsoft Edge.
|
bug,Navigation
|
low
|
Minor
|
2,677,200,814 |
ollama
|
Request: Nexa AI Omnivision
|
Here is the link to Nexa AI's model. (I haven't checked over it to make sure it's super reputable though)
https://huggingface.co/NexaAIDev/omnivision-968M
|
model request
|
low
|
Major
|
2,677,208,897 |
ant-design
|
`RangePicker` breaks when using with date-fns via `dateFnsGenerateConfig`
|
### Reproduction link
[](https://codesandbox.io/p/sandbox/datepicker-rangepicker-antd-v5-hr2rr9?file=%2Findex.tsx)
### Steps to reproduce
- Setup Antd DatePicker to use date-fns according the [official guide](https://ant.design/docs/react/use-custom-date-library#use-date-fns);
- Renders `RangePicker` using `format`, `picker` containing `"time"` value and `use12Hours` props;
- Click on the RangePicker to input values.
### What is expected?
The RangePicker should show the dropdown with the options so we can select a time range.
### What is actually happening?
The entire app breaks with an error "Format string contains an unescaped latin alphabet character `A`"
| Environment | Info |
| --- | --- |
| antd | 5.22.1 |
| React | ^18.0.0 |
| System | macOS Sequoia Version 15.0 (24A335) |
| Browser | Google Chrome Version 131.0.6778.86 |
---
I walked through the previous versions to find when this starts to break and I found the `v5.18.3` working but if we upgrade to `v5.19.0` or earlier, the error starts to happen.
<!-- generated by ant-design-issue-helper. DO NOT REMOVE -->
|
help wanted,Inactive
|
low
|
Critical
|
2,677,214,240 |
ui
|
[bug]: GitHub link in site header is only clickable on icon
|
### Describe the bug
The GitHub link in the site header is only clickable on the icon, if the button is clicked outside of the icon, the link doesn't work.
This is solved by using asChild as documented in https://ui.shadcn.com/docs/components/button
### Affected component/components
Button, Link
### How to reproduce
1. Go to https://ui.shadcn.com/
2. Click on the GitHub link in the site header, clicking outside of the icon
3. Doesn't work
### Codesandbox/StackBlitz link
_No response_
### Logs
_No response_
### System Info
```bash
Any system
```
### Before submitting
- [X] I've made research efforts and searched the documentation
- [X] I've searched for existing issues
|
bug
|
low
|
Critical
|
2,677,214,306 |
flutter
|
Removing the `Key` class
|
### Document Link
[flutter.dev/go/unwrapping-widget-keys](https://docs.google.com/document/d/1Fc9ZYTTxlJUAfGeQzf83wXuV-wSiRZFHhMgwoUNukhI)
### What problem are you solving?
This is an effort to make widgets less complicated and more performant.
|
c: API break,a: first hour,c: tech-debt,design doc,:scroll:
|
low
|
Minor
|
2,677,221,233 |
neovim
|
Windows WSL: osc52 data with CRLF lines gets Chinese chars
|
### Problem
osc52 works for one line only. More than one line means bug (random chinese character etc.)
### Steps to reproduce
nvim --clean -u minimal.lua
write
deed
de
select previous text
press "y"
your clipboard now is chinese
ๆคๆฅๆe
note on my full config (not just the minimal.lua) copying a selection always copy only the first line instead of becoming chinese clipboard.
Also copying one line only always works flawlessly
minimal.lua
```lua
vim.g.clipboard = {
name = 'OSC 52',
copy = {
['+'] = require('vim.ui.clipboard.osc52').copy('+'),
['*'] = require('vim.ui.clipboard.osc52').copy('*'),
},
paste = {
['+'] = require('vim.ui.clipboard.osc52').paste('+'),
['*'] = require('vim.ui.clipboard.osc52').paste('*'),
},
}
vim.cmd("set clipboard=unnamedplus")
```
### Expected behavior
Not chinese characters
### Nvim version (nvim -v)
0.10.2
### Vim (not Nvim) behaves the same?
vim no osc52 native AFAIK
### Operating system/version
windows wsl ubuntu
### Terminal name/version
windows terminal
### $TERM environment variable
on ssh xterm-256color
### Installation
appimage 0.10.2 github
|
bug,platform:windows,clipboard,needs:repro,platform:wsl
|
low
|
Critical
|
2,677,333,920 |
node
|
Support `concurrency` when `--experimental-test-isolation` is set to `'none'`
|
### What is the problem this feature will solve?
It would be nice to be able to run multiple "workers" when using the `none` test isolation mode.
Today, according to the docs, when `--experimental-test-isolation` is set to `'none'`, it implies 1 concurrency: https://github.com/nodejs/node/blob/f270462c09ddfd770291a7c8a2cd204b2c63d730/doc/api/cli.md#L2252-L2264
However, `mocha` another test runner that does not isolate tests, does accept a concurrency flag: https://mochajs.org/#parallel-tests
### What is the feature you are proposing to solve the problem?
We're looking to move off of `jest` because its test module isolation is extremely slow. We love the idea of using the node-native test runner with isolation disabled, instead of adopting another third-party framework like `mocha`.
However, we'd need to write some custom code (e.g., using [`parallel`](https://www.gnu.org/software/parallel/)) to spin up `n` concurrent, isolation-disabled tests to effectively utilize all the cores available on our CI machine.
The docs and other recent comments all indicate that when `--experimental-test-isolation` is set to `'none'`, concurrency must be `1`. However, I couldn't find the reasoning in [the original PR](https://github.com/nodejs/node/pull/53927) or [issue](https://github.com/nodejs/node/issues/51548).
There's probably a good reason for this but, as someone not intimately familiar with the implementation, it's not obvious to me why we wouldn't be able to run non-isolated tests concurrently, like in `mocha`.
### What alternatives have you considered?
I could probably use a tool like [`parallel`](https://www.gnu.org/software/parallel/) to spin up multiple calls to `node --test`. However, this would require me to also write code to split up all the test files between the parallel runs, etc.
|
feature request,test_runner
|
low
|
Major
|
2,677,337,079 |
pytorch
|
cudagraph_trees error message suggestion doesn't always work
|
This came from running a user empathy model. Repro is:
```
import torch
class Mod(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, x):
out = x * 2
if not hasattr(self, "t"):
self.t = torch.arange(x.shape[1], device='cuda')[None, None]
elif self.t.shape[-1] < x.shape[1]:
self.t = torch.arange(x.shape[1], device='cuda')[None, None]
else:
self.t = self.t[..., :x.shape[1]]
return out
m = Mod()
m = torch.compile(m, mode='reduce-overhead')
x = torch.randn(8, 8, device='cuda')
with torch.no_grad():
torch.compiler.cudagraph_mark_step_begin()
m(x)
torch.compiler.cudagraph_mark_step_begin()
m(x)
torch.compiler.cudagraph_mark_step_begin()
m(x)
```
This gives the error:
```
File "/home/hirsheybar/.conda/envs/pytorch_nightly/lib/python3.11/site-packages/torch/_inductor/cudagraph_trees.py", line 628, in get_non_cudagraph_inps
and t.untyped_storage().data_ptr() not in existing_path_data_ptrs
^^^^^^^^^^^^^^^^^^^
RuntimeError: Error: accessing tensor output of CUDAGraphs that has been overwritten by a subsequent run. Stack trace: File "/data/users/hirsheybar/nightly/evo/tmp2.py", line 10, in forward
self.t = torch.arange(x.shape[1], device='cuda')[None, None]. To prevent overwriting, clone the tensor outside of torch.compile() or call torch.compiler.cudagraph_mark_step_begin() before each model invocation.
```
cc @mcarilli @ezyang @eellison @penguinwu @chauhang @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @yf225 @chenyang78 @kadeng @muchulee8 @ColinPeppler @amjames @desertfire @aakhundov
|
triaged,module: cuda graphs,oncall: pt2,module: inductor,empathy-day
|
low
|
Critical
|
2,677,359,945 |
excalidraw
|
[Feature Request] - Add a document to the canvas
|
I think a very good addition to Excalidraw would be to add a "Document" feature.
Either the way Whimsical does it, where you can create documents separated from canvas, or the way Eraser does it, with a document attached to the canvas itself.
|
enhancement
|
low
|
Minor
|
2,677,368,559 |
angular
|
Can't build adev (bazel error)
|
### Which @angular/* package(s) are the source of the bug?
Don't known / other
### Is this a regression?
Yes
### Description
I get this weird error from bazel when I try build adev
### Please provide a link to a minimal reproduction of the bug
I followed every step here https://github.com/angular/angular/blob/main/adev/README.md, also tried cleaning after error but did not help
### Please provide the exception or error you saw
```true
ERROR: C:/users/cros4/desktop/angular/packages/zone.js/BUILD.bazel:19:8: Executing genrule //packages/zone.js:LICENSE_wrapped failed: (Exit -1): bash.exe failed: error executing command c:\tools\msys64\usr\bin\bash.exe -c ... (remaining 1 argument skipped)
Action failed to execute: java.io.IOException: ERROR: src/main/native/windows/process.cc(202): CreateProcessW("c:\tools\msys64\usr\bin\bash.exe" -c "source external/bazel_tools/tools/genrule/genrule-setup.sh; (echo '/**
@license' && cat LICENSE && echo '*/') > bazel-out/x64_windows-fastbuild/bin/packages/zone.js/LICENSE.wrapped"): Can't find specified file.
(error: 2)
Target //adev:serve failed to build
```
### Please provide the environment you discovered this bug in (run `ng version`)
```true
OS: win32 x64
```
### Anything else?
_No response_
|
area: docs-infra
|
low
|
Critical
|
2,677,369,816 |
pytorch
|
`CrossEntropyLoss` much slower with class probabilities vs. class indices target
|
### ๐ Describe the bug
Hi, I noticed this when training ViT-S/16 on the ImageNet-1k dataset, with and without [MixUp](https://arxiv.org/abs/1710.09412). In short: torchvision's [`v2.MixUp`](https://pytorch.org/vision/main/generated/torchvision.transforms.v2.MixUp.html), just like other common implementations, converts the class indices to one-hot class probabilities before linearly interpolate them, so the same call to `CrossEntropyLoss()` goes to different kernels with vs. without MixUp. In a less optimized setting ([specific commit](https://github.com/EIFY/mup-vit/tree/51bcdc77ea6e26ceef049e96f59702a9d9ef15d1)), I found that with vs. without MixUp makes the difference between **0.198s vs. 0.120s** per step (excluding data loading) with batch size 1024 on a 8x A100-SXM4-40GB [Lambda](https://lambdalabs.com/) instance. In further iterations, I included `CrossEntropyLoss` in the model so they compile together. I then tested a workaround in which the custom MixUp returns both targets and the weight scalar lam(bda):
```python
lam = self._dist.sample()
images = images.roll(1, 0).mul_(1.0 - lam).add_(images, alpha=lam)
return images, lam, labels, labels.roll(1, 0)
```
and the model calculates the cross entropy loss by using the class indices twice:
```python
logprob = F.log_softmax(out, dim=1)
return lam * F.nll_loss(logprob, target1) + (1.0 - lam) * F.nll_loss(logprob, target2)
```
See https://github.com/EIFY/mup-vit/compare/normal-prefetch...loss-optim. I found that this is faster (**0.184s** per step) than just calling `CrossEntropyLoss()` with class probabilities even with `torch.compile(...,mode="max-autotune-no-cudagraphs")` (**0.191s** per step), excluding data loading. Is this expected? Intuitively I can see that the sparsity helps forward & backward pass in loss calculation, but backward pass beyond that should be the same and it's hard to understand the observed second/step difference with a 12-layer transformer.
To replicate the timing result, you can run the code with fake data:
```shell
MUPVIT_MAIN=~/Downloads/mup-vit/main.py
PYTHON=torchrun
N_WORKERS=80
N_THREADS=124
EPOCH=90
NUMEXPR_MAX_THREADS=$N_THREADS $PYTHON $MUPVIT_MAIN --fake-data --workers $N_WORKERS --multiprocessing-distributed --epochs $EPOCH --batch-size 1024 --mlp-head --report-to wandb --name cross-entropy-loss-class-prob
NUMEXPR_MAX_THREADS=$N_THREADS $PYTHON $MUPVIT_MAIN --fake-data --workers $N_WORKERS --multiprocessing-distributed --epochs $EPOCH --batch-size 1024 --mlp-head --mixup-alpha 0.0 --report-to wandb --name cross-entropy-loss-class-index
```
with either the [normal-prefetch](https://github.com/EIFY/mup-vit/tree/normal-prefetch) or the [loss-optim](https://github.com/EIFY/mup-vit/tree/loss-optim) branch.
### Versions
The output below shows PyTorch version: 2.3.1 and torchvision==0.18.1, but I am actually running
```
$ pip freeze
(...)
torch==2.5.1
torchaudio==2.5.1
torchvision==0.20.1
(...)
```
I don't know why `python3 -mpip list --format=freeze` finds the old packages.
```
Collecting environment information...
PyTorch version: 2.3.1
Is debug build: False
CUDA used to build PyTorch: 12.4
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.5 LTS (x86_64)
GCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
Clang version: Could not collect
CMake version: version 3.22.1
Libc version: glibc-2.35
Python version: 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0] (64-bit runtime)
Python platform: Linux-6.8.0-47-generic-x86_64-with-glibc2.35
Is CUDA available: True
CUDA runtime version: 12.4.131
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration:
GPU 0: NVIDIA A100-SXM4-40GB
GPU 1: NVIDIA A100-SXM4-40GB
GPU 2: NVIDIA A100-SXM4-40GB
GPU 3: NVIDIA A100-SXM4-40GB
GPU 4: NVIDIA A100-SXM4-40GB
GPU 5: NVIDIA A100-SXM4-40GB
GPU 6: NVIDIA A100-SXM4-40GB
GPU 7: NVIDIA A100-SXM4-40GB
Nvidia driver version: 550.90.12
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 48 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 124
On-line CPU(s) list: 0-123
Vendor ID: AuthenticAMD
Model name: AMD EPYC 7542 32-Core Processor
CPU family: 23
Model: 49
Thread(s) per core: 1
Core(s) per socket: 1
Socket(s): 124
Stepping: 0
BogoMIPS: 5800.00
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities
Virtualization: AMD-V
Hypervisor vendor: KVM
Virtualization type: full
L1d cache: 7.8 MiB (124 instances)
L1i cache: 7.8 MiB (124 instances)
L2 cache: 62 MiB (124 instances)
L3 cache: 1.9 GiB (124 instances)
NUMA node(s): 2
NUMA node0 CPU(s): 0-31,64-95
NUMA node1 CPU(s): 32-63,96-123
Vulnerability Gather data sampling: Not affected
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Reg file data sampling: Not affected
Vulnerability Retbleed: Mitigation; untrained return thunk; SMT disabled
Vulnerability Spec rstack overflow: Vulnerable: Safe RET, no microcode
Vulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] flake8==4.0.1
[pip3] numpy==1.21.5
[pip3] optree==0.12.1
[pip3] torch==2.3.1
[pip3] torchvision==0.18.1
[pip3] triton==2.3.1
[conda] Could not collect
```
cc @albanD @mruberry @jbschlosser @walterddr @mikaylagawarecki
|
module: nn,module: loss,triaged
|
low
|
Critical
|
2,677,437,085 |
go
|
proposal: cmd/{go,cgo,compile}: change interpretation of relative //line directives
|
**Background**: generated Go source files may contain //line directives that relate each line of Go source to the line of some other file from which it was generated. If the file named by the line directive is a relative path, cmd/compile currently interprets it like any other relative path, that is, relative to its own os.Getwd. This demands that the tool that generated the //line directive has the same working directory as the compiler (e.g. GOROOT/src), or that it creates the line directive in anticipation of the consuming process having (say) GOROOT/src as its working directory.
By contrast, go/scanner takes a different approach: it interprets relative line directives relative to the scanned file's directory. So, if it encounters the line directive `//line a.yacc:1` in a file named `a/a.go`, the relative line directive is resolved as `a/a.yacc`. (In March 2018 go/scanner was changed to resemble the compiler in [CL 100235](https://go.dev/cl/100235), but this was reverted in August 2018 by [CL 127658](https://go.dev/cl/127658).)
The difference in behavior is easily observed by introducing a dummy line directive with a compiler error into an arbitrary Go source file. In this case, we'll append a relative reference from gopls/main.go to its README.md file:
```diff
xtools$ git diff
diff --git a/gopls/main.go b/gopls/main.go
index 083c4efd8..b63784ba8 100644
--- a/gopls/main.go
+++ b/gopls/main.go
@@ -34,3 +34,6 @@ func main() {
ctx := context.Background()
tool.Main(ctx, cmd.New(), os.Args[1:])
}
+
+//line README.md:1
+invalid
```
When we build the package, the error message names a non-existent file:
```
xtools$ go build -o /dev/null ./gopls
# golang.org/x/tools/gopls
README.md:1: syntax error: non-declaration statement outside function body
```
When we use tools based on go/scanner, the error message names the correct file:
```
xtools$ gopackages -mode=allsyntax -deps ./gopls 2>&1 | grep README
/Users/adonovan/w/xtools/gopls/README.md:1: expected declaration, found invalid
```
This is true even when we run the same commands from a different directory. Again, `go build` names the non-existent file, and go/scanner names the correct file:
```
xtools$ cd internal/
internal$ go build -o /dev/null ../gopls
# golang.org/x/tools/gopls
README.md:1: syntax error: non-declaration statement outside function body
```
```
internal$ gopackages -mode=allsyntax -deps ../gopls 2>&1 | grep README
/Users/adonovan/w/xtools/gopls/README.md:1: expected declaration, found invalid
```
**Proposal:** The compiler's scanner should interpret relative line directives relative to the directory of the scanned file, and this behavior should be documented at [https://pkg.go.dev/cmd/compile#hdr-Compiler_Directives](https://pkg.go.dev/cmd/compile#hdr-Compiler_Directives).
This may require changes to the line directives emitted by cgo and other tools (including [unparam and unconvert](https://go.dev/cl/127658), whatever they are).
Arguably, this is an incompatible change, but we notice that more often than not, when published Go modules use relative line directives, they already follow the proposed discipline.
See also:
- [https://github.com/golang/go/issues/24183](https://github.com/golang/go/issues/24183)
- [https://github.com/golang/go/issues/26671](https://github.com/golang/go/issues/26671)
- [https://github.com/golang/go/issues/69689](https://github.com/golang/go/issues/69689)
@griesemer @ianlancetaylor @findleyr @timothy-king
|
Proposal
|
low
|
Critical
|
2,677,461,467 |
flutter
|
[Android] Dragging outside the text should not show the magnifier
|
### Steps to reproduce
1. Run an app with a text field (such as given below) on an Android device.
2. Enter some text.
3. Perform a drag inside the field but not on any of the text.
### Expected results
The magnifier hides whenever the drag leaves the text.
### Actual results
The magnifier is shown during the whole drag.
### Code sample
<details><summary>Code sample</summary>
```dart
import 'package:flutter/material.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'Flutter Demo',
theme: ThemeData(
primarySwatch: Colors.blue,
),
home: const MyHomePage(title: 'Flutter Demo Home Page'),
);
}
}
class MyHomePage extends StatelessWidget {
const MyHomePage({super.key, required this.title});
final String title;
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(
title: Text(title),
),
body: Center(
child: Column(
mainAxisAlignment: MainAxisAlignment.center,
children: const <Widget>[
TextField(
),
],
),
),
);
}
}
```
</details>
### Screenshots or Video
<details open>
<summary>Video of native Android</summary>
https://github.com/user-attachments/assets/03f2f56a-25ec-47f7-be21-26b9037c4be1
</details>
<details open>
<summary>Video of Flutter</summary>
https://github.com/user-attachments/assets/a295f2ee-5978-400a-a09f-97fd4760ebc6
</details>
Discovered while working on https://github.com/flutter/flutter/pull/148784.
### Flutter Doctor output
<details><summary>Doctor output</summary>
```console
flutter doctor
Building flutter tool...
Resolving dependencies... (1.0s)
Downloading packages...
Got dependencies.
Doctor summary (to see all details, run flutter doctor -v):
[!] Flutter (Channel [user-branch], 3.22.0-35.0.pre.2441, on Debian GNU/Linux rodete 6.9.10-1rodete5-amd64, locale en_US.UTF-8)
! Flutter version 3.22.0-35.0.pre.2441 on channel [user-branch] at /usr/local/google/home/jmccandless/Projects/flutter
Currently on an unknown channel. Run `flutter channel` to switch to an official channel.
If that doesn't fix the issue, reinstall Flutter by following instructions at https://flutter.dev/setup.
[โ] Android toolchain - develop for Android devices (Android SDK version 35.0.0)
[โ] Chrome - develop for the web
[โ] Linux toolchain - develop for Linux desktop
[โ] Android Studio (version 2023.3)
[โ] Connected device (3 available)
[โ] Network resources
! Doctor found issues in 1 category.
```
</details>
|
platform-android,P2,team-text-input,triaged-text-input
|
low
|
Minor
|
2,677,551,234 |
opencv
|
Assertion error when flipping arrays with more than 512 channels
|
### System Information
OpenCV python version: 4.10.0.84
Operating System / Platform: MacOS 15.1.1
Python version: 3.12.7
### Detailed description
Flipping arrays with more than 512 channels (third dimension) throws the following assertion error:
```
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
cv2.error: OpenCV(4.10.0) /Users/xperience/GHA-Actions-OpenCV/_work/opencv-python/opencv-python/opencv/modules/core/src/matrix_transform.cpp:784: error: (-215:Assertion failed) _src.dims() <= 2 in function 'flip'
```
### Steps to reproduce
```python
import cv2
import numpy as np
# Works
arr1 = np.zeros((256, 256, 512), dtype=np.uint8)
cv2.flip(arr1, 0)
# Throws error: (-215:Assertion failed)
arr2 = np.zeros((256, 256, 513), dtype=np.uint8)
cv2.flip(arr2, 0)
```
### Issue submission checklist
- [X] I report the issue, it's not a question
- [X] I checked the problem with documentation, FAQ, open issues, forum.opencv.org, Stack Overflow, etc and have not found any solution
- [X] I updated to the latest OpenCV version and the issue is still there
- [X] There is reproducer code and related data files (videos, images, onnx, etc)
|
bug,category: python bindings
|
low
|
Critical
|
2,677,551,526 |
rust
|
error: rustc interrupted by SIGSEGV
|
<!--
Thank you for filing a bug report! ๐ Please provide a short summary of the bug,
along with any information you feel relevant to replicating the bug.
-->
I tried this code: https://github.com/warcayac/axum_postgre_email
```rust
cargo run
```
I expected to see this happen: *explanation*
```
โ
Connection to the database is successful!
๐ Listening on 127.0.0.1:8080
```
Instead, this happened: *explanation*
```
error: rustc interrupted by SIGSEGV, printing backtrace
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x33b6d83) [0x7a483d9b6d83]
/usr/lib/libc.so.6(+0x3d1d0) [0x7a483a4171d0]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCs9f8yWUR3ijp_10rustc_lint7contextNtB4_9LintStore15check_lint_name+0x69b) [0x7a483f5cf0db]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4fd040d) [0x7a483f5d040d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9f8yWUR3ijp_10rustc_lint6levels22shallow_lint_levels_on+0x303) [0x7a483f8acb0d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x52ac7aa) [0x7a483f8ac7aa]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4c1d5) [0x7a483f14c1d5]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4be0d) [0x7a483f14be0d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCshbMUE98pR8y_12rustc_middle4lintNtB4_19ShallowLintLevelMap21lint_level_id_at_node+0x1757) [0x7a483f5dbb17]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor11visit_block+0x24b) [0x7a483f5d694b]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafety14check_unsafety+0x4ca) [0x7a483c19e84a]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513ff97) [0x7a483f73ff97]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x51401ca) [0x7a483f7401ca]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513fec1) [0x7a483f73fec1]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5449b0f) [0x7a483fa49b0f]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9EojWPpxxi0_15rustc_interface6passes8analysis+0x19) [0x7a483fa41da5]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5441d89) [0x7a483fa41d89]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6d22) [0x7a483fec6d22]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6acf) [0x7a483fec6acf]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x57860b2) [0x7a483fd860b2]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x583ed16) [0x7a483fe3ed16]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58779b0) [0x7a483fe779b0]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5877d2b) [0x7a483fe77d2b]
/usr/lib/libc.so.6(+0x9439d) [0x7a483a46e39d]
/usr/lib/libc.so.6(+0x11949c) [0x7a483a4f349c]
note: we would appreciate a report at https://github.com/rust-lang/rust
help: you can increase rustc's stack size by setting RUST_MIN_STACK=16777216
note: backtrace dumped due to SIGSEGV! resuming signal
Finished `dev` profile [unoptimized + debuginfo] target(s) in 0.23s
Running `target/debug/postgresql_email_backend`
โ
Connection to the database is successful!
๐ Listening on 127.0.0.1:8080
```
After that I tried with: `RUST_MIN_STACK=16777216 cargo run`, same result
### Meta
<!--
If you're using the stable version of the compiler, you should also check if the
bug also exists in the beta or nightly versions.
-->
`rustc --version --verbose`:
```
rustc 1.82.0 (f6e511eec 2024-10-15)
binary: rustc
commit-hash: f6e511eec7342f59a25f7c0534f1dbea00d01b14
commit-date: 2024-10-15
host: x86_64-unknown-linux-gnu
release: 1.82.0
LLVM version: 19.1.1
```
<!--
Include a backtrace in the code block by setting `RUST_BACKTRACE=1` in your
environment. E.g. `RUST_BACKTRACE=1 cargo build`.
-->
<details><summary>Backtrace</summary>
<p>
```
RUST_BACKTRACE=1 cargo run
error: rustc interrupted by SIGSEGV, printing backtrace
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x33b6d83) [0x7a483d9b6d83]
/usr/lib/libc.so.6(+0x3d1d0) [0x7a483a4171d0]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCs9f8yWUR3ijp_10rustc_lint7contextNtB4_9LintStore15check_lint_name+0x69b) [0x7a483f5cf0db]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4fd040d) [0x7a483f5d040d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9f8yWUR3ijp_10rustc_lint6levels22shallow_lint_levels_on+0x303) [0x7a483f8acb0d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x52ac7aa) [0x7a483f8ac7aa]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4c1d5) [0x7a483f14c1d5]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x4b4be0d) [0x7a483f14be0d]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvMs_NtCshbMUE98pR8y_12rustc_middle4lintNtB4_19ShallowLintLevelMap21lint_level_id_at_node+0x1757) [0x7a483f5dbb17]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor11visit_block+0x24b) [0x7a483f5d694b]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0x15e3) [0x7a483c1a4da3]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvXs1_NtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafetyNtB5_15UnsafetyVisitorNtNtNtCshbMUE98pR8y_12rustc_middle4thir5visit7Visitor10visit_expr+0xf8) [0x7a483c1a38b8]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs5Tw7peOzaM9_15rustc_mir_build14check_unsafety14check_unsafety+0x4ca) [0x7a483c19e84a]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513ff97) [0x7a483f73ff97]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x51401ca) [0x7a483f7401ca]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x513fec1) [0x7a483f73fec1]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5449b0f) [0x7a483fa49b0f]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(_RNvNtCs9EojWPpxxi0_15rustc_interface6passes8analysis+0x19) [0x7a483fa41da5]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5441d89) [0x7a483fa41d89]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6d22) [0x7a483fec6d22]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58c6acf) [0x7a483fec6acf]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x57860b2) [0x7a483fd860b2]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x583ed16) [0x7a483fe3ed16]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x58779b0) [0x7a483fe779b0]
/home/warcayac/.rustup/toolchains/stable-x86_64-unknown-linux-gnu/lib/librustc_driver-2d2db9d3929202cf.so(+0x5877d2b) [0x7a483fe77d2b]
/usr/lib/libc.so.6(+0x9439d) [0x7a483a46e39d]
/usr/lib/libc.so.6(+0x11949c) [0x7a483a4f349c]
note: we would appreciate a report at https://github.com/rust-lang/rust
help: you can increase rustc's stack size by setting RUST_MIN_STACK=16777216
note: backtrace dumped due to SIGSEGV! resuming signal
Finished `dev` profile [unoptimized + debuginfo] target(s) in 0.21s
Running `target/debug/postgresql_email_backend`
โ
Connection to the database is successful!
๐ Listening on 127.0.0.1:8080
^C
```
</p>
</details>
|
I-crash,T-compiler,A-incr-comp,C-bug,S-needs-repro
|
low
|
Critical
|
2,677,577,476 |
pytorch
|
Dynamo graph break on RNN/LSTM doesn't resume the function correctly.
|
### ๐ Describe the bug
Dynamo purposely graph breaks on RNN, GRU, LSTMs. We expect the sub modules before and after the RNN/GRU/LSTM calls should be compiled as partial graphs, however, we found only the sub modules before the graph break got compiled, the one after doesn't.
Repro:
```
import torch
import torch.nn as nn
import torch.nn.functional as F
class MultiLayerModel(nn.Module):
def __init__(self, input_dim, lstm_hidden_dim, conv_channels, linear_out_dim, num_classes):
super(MultiLayerModel, self).__init__()
# Convolutional layer
self.conv = nn.Conv2d(in_channels=1, out_channels=conv_channels, kernel_size=3, stride=1, padding=1)
# Layer normalization for convolution output
self.conv_layer_norm = nn.LayerNorm([conv_channels, input_dim, input_dim])
# LSTM layer
self.lstm = nn.LSTM(input_dim, lstm_hidden_dim, batch_first=True, bidirectional=True)
# Fully connected linear layers
self.fc1 = nn.Linear(lstm_hidden_dim * 2, linear_out_dim) # LSTM is bidirectional, so hidden_dim * 2
self.fc2 = nn.Linear(linear_out_dim, num_classes)
# Layer normalization for linear layers
self.linear_layer_norm = nn.LayerNorm(linear_out_dim)
def forward(self, x):
# Convolutional layer expects 4D input: (batch_size, channels, height, width)
x = self.conv(x)
x = F.relu(x)
x = self.conv_layer_norm(x)
# Flatten spatial dimensions for the LSTM (assuming square input images)
batch_size, channels, height, width = x.size()
x = x.view(batch_size, channels * height, width) # Prepare for LSTM: (batch, seq_len, features)
# LSTM expects input: (batch_size, seq_len, input_dim)
x, _ = self.lstm(x)
# Fully connected layers
x = F.relu(self.fc1(x[:, -1, :])) # Use the output from the last time step
x = self.linear_layer_norm(x)
x = self.fc2(x)
return x
input_dim = 28
lstm_hidden_dim = 128
conv_channels = 16
linear_out_dim = 64
num_classes = 10
model = MultiLayerModel(input_dim, lstm_hidden_dim, conv_channels, linear_out_dim, num_classes)
print(model)
model = torch.compile(model)
# Dummy input: Batch of grayscale images (batch_size, channels, height, width)
dummy_input = torch.randn(8, 1, 28, 28)
output = model(dummy_input)
print("Output shape:", output.shape)
```
Run this script with ```TORCH_LOGS=+dynamo```, you will find there is only one compiled graph which includes the sub module befroe line ```x = x.view(batch_size, channels * height, width)```.
### Versions
N/A
cc @chauhang @penguinwu @voznesenskym @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @kadeng @amjames
|
triaged,bug,oncall: pt2,module: dynamo
|
low
|
Critical
|
2,677,578,157 |
go
|
build: build failure on go1.23-linux-amd64_c3h88-perf_vs_release
|
```
#!watchflakes
default <- builder == "go1.23-linux-amd64_c3h88-perf_vs_release" && repo == "go" && mode == "build"
```
Issue created automatically to collect these failures.
Example ([log](https://ci.chromium.org/b/8730818689678245169)):
go: downloading github.com/BurntSushi/toml v1.0.0
go: downloading github.com/kballard/go-shellquote v0.0.0-20180428030007-95032a82bc51
2024/11/19 18:09:12 Load average: 0.67 0.20 0.07 1/1176 1834
2024/11/19 18:09:12 Waiting for load average to drop below 0.20...
2024/11/19 18:09:42 Load average: 0.47 0.19 0.07 1/1177 1834
2024/11/19 18:09:42 Waiting for load average to drop below 0.20...
2024/11/19 18:10:12 Load average: 0.28 0.17 0.07 2/1177 1834
2024/11/19 18:10:12 Waiting for load average to drop below 0.20...
2024/11/19 18:10:42 Load average: 0.23 0.17 0.07 3/1180 1834
2024/11/19 18:10:42 Waiting for load average to drop below 0.20...
...
[sweet] Running benchmark tile38 for experiment: run 8
[sweet] Running benchmark tile38 for baseline: run 8
[sweet] Running benchmark tile38 for experiment: run 9
[sweet] Running benchmark tile38 for baseline: run 9
[sweet] Running benchmark tile38 for experiment: run 10
[sweet] Running benchmark tile38 for baseline: run 10
[sweet] error: failed to execute benchmarks: gvisor
2024/11/19 23:45:15 Error running sweet: error running sweet run: exit status 1
2024/11/19 23:45:15 FAIL
exit status 1
โ [watchflakes](https://go.dev/wiki/Watchflakes)
|
NeedsInvestigation
|
low
|
Critical
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.