text
stringlengths
20
57.3k
labels
class label
4 classes
Title: CI test linux://python/ray/data:test_numpy_support is flaky Body: CI test **linux://python/ray/data:test_numpy_support** is consistently_failing. Recent failures: - https://buildkite.com/ray-project/postmerge/builds/8496#01952c44-0d09-482f-a6f3-aa6493261d16 - https://buildkite.com/ray-project/postmerge/builds/8495#01952b53-0f7b-45fe-9a74-7107914282da - https://buildkite.com/ray-project/postmerge/builds/8495#01952b30-22c6-4a0f-9857-59a7988f67d8 - https://buildkite.com/ray-project/postmerge/builds/8491#01952b00-e020-4d4e-b46a-209c0b3dbf5b - https://buildkite.com/ray-project/postmerge/builds/8491#01952ad9-1225-449b-84d0-29cfcc6a048c DataCaseName-linux://python/ray/data:test_numpy_support-END Managed by OSS Test Policy
1medium
Title: `edit_polygons` unable to transform CRS correctly Body: <!-- Please search existing issues to avoid creating duplicates. --> ### Environment Information - leafmap version: 0.38.8 - Python version: 3.10.12 - Operating System: Ubuntu 20.04.6 LTS ### Description This issue is similar to #920. `edit_polygons` does not work if `gdf` is not in EPSG:4326. ### What I Did ```py show_what_fails = True import leafmap.leafmap as leafmap import geopandas as gpd m = leafmap.Map() m.add_basemap("HYBRID") gdf = gpd.read_file("https://raw.githubusercontent.com/opengeos/leafmap/master/examples/data/countries.geojson") if show_what_fails: gdf = gdf.to_crs("EPSG:3857") else: gdf = gdf.to_crs("EPSG:4326") m.edit_polygons(gdf) m ```
1medium
Title: 反应个问题 Body: 语言云的在线演示功能,经常性的崩溃,想看个例子都不行,好几天了都是这样,这两个月不知道碰到多少种这种情况了。希望及时处理! ![123](https://user-images.githubusercontent.com/20137919/29362941-86ddac22-82c0-11e7-8739-4862065a3fb1.png)
1medium
Title: [Nodriver] Features Body: Consider adding to nodriver some features: 1. For `Element` implement same methods like for tab (`select`, `select_all`, if possible also `find`) 2. For `send_keys` add cdp commands like `keyDown`, `keyUp`. It will mimic events from real keyboard as they are not fired when simply sending text 3. From selenium driverless feature `human mouse move`. It would be good probably to mimic human actions on the website by not simply clicking button, but also moving mouse to it 4. Mimic `human scrolling` (like in intervals with speed by normal distribution). It can probably replace scrolling into view better 5. Some method `switch_to_iframe` to switch to iframe (return `Tab`). It will be quite good because without this option `--disable-web-security`, iframe is created as explicit tab which can be found in targets of driver. 6. !!! Add timeout for cdp commands on send. For some reason when I try to edit outcoming request, after sending some cdp commands, driver will not respond on target updating. Like there are no response probably from the chrome, but it also indefinitely blocks code execution and does not throw anything. 7. (optional) For `Transaction` when it is not completed, property `has_exception` always returns `True` because there are still no value I will extend this in future probably.
1medium
Title: Move themes out to separate files Body: All current themes are defined inside a `theme.py` module. It looks like the next: ``` elif theme == "midnight": # {{{ midnight # Based on XCode's midnight theme # Looks best in a console with green text against black background palette_dict.update({ "variables": ("white", "default"), "var label": ("light blue", "default"), "var value": ("white", "default"), ... ``` I propose to create a separate folder, for example, `themes`. Each theme will be as a separate file JSON or YAML inside a created folder. I think the theme is not related to an application. Consider a terminal or popular text editors. All of their themes are separate files which easy to install without changing a code.
1medium
Title: Unexpected Output Channel Reduction in skimage.transform.rescale without Explicit channel_axis Body: ### Description: When using the `skimage.transform.rescale` function to downscale an RGB image without explicitly specifying the `channel_axis`, the output image unexpectedly has two channels instead of the standard three. This is observed when applying the function to the standard RGB image from `skimage.data.astronaut()`. The expected output shape for a rescaled RGB image is `(height, width, 3)`. However, the observed output shape is `(256, 256, 2)`. Since `channel_axis` is an optional parameter, the default behavior should maintain the number of channels in the input image. ### Way to reproduce: ``` import numpy as np from skimage import transform, data # Load an RGB image rgb_image = data.astronaut() # Rescale the image rescaled_image = transform.rescale(rgb_image, 0.5) # Output the shape print(rescaled_image.shape) ``` #### Expected Behavior: The output shape should be `(256, 256, 3)` for a correctly rescaled RGB image, preserving the number of channels. #### Observed Behavior: The output shape is `(256, 256, 2)`, indicating a reduction in the number of channels. ### Version information: ```Shell 3.9.6 (default, Mar 10 2023, 20:16:38) [Clang 14.0.3 (clang-1403.0.22.14.1)] macOS-13.3.1-x86_64-i386-64bit scikit-image version: 0.22.0 numpy version: 1.26.2 ```
2hard
Title: Missing dependencies between tasks if Airflow Task Group is empty Body: ### Apache Airflow version 2.10.5 ### If "Other Airflow 2 version" selected, which one? _No response_ ### What happened? Airflow is not honouring the task dependecies for a task if the parent is empty task group ### What you think should happen instead? If the parent of a task is empty task group Then it should follow the dependency rule of the parent's parent. ### How to reproduce Define a dag with 5 tasks A B C D E C D belongs to task group test Dependencies between tasks is A >> B >> test >> E C >> D If task group test is not empty, everything is running fine If task group test is empty Ideally behaviour should be A >> B>> E But tasks A, E are running parallely without honouring the dependencies Graph view is showing the dependencies correctly but execution is being different ### Operating System Linux ### Versions of Apache Airflow Providers _No response_ ### Deployment Other ### Deployment details _No response_ ### Anything else? _No response_ ### Are you willing to submit PR? - [x] Yes I am willing to submit a PR! ### Code of Conduct - [x] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
1medium
Title: Add zip join to combine fixtures Body: Consider this issue as a feature request please. I suggest to add more flexibility to how parameters for tests are combined. In pytest tests are executed for each combination (cartesian product) of all parameters. I was glad to find fixture_union in pytest cases that introduces flexibility in combining fixture parameters for tests. I suggest to add further options to combine fixtures: zip join to combine fixture sets by position number in a set. This will allow to test actual and expected states determined for each test object in different fixtures: ``` @pytest_fixture_plus(params=[obj1, obj2, obj3]) def actual_state(request): return produce_actual_state(request.param) @pytest_fixture_plus(params=[obj1, obj2, obj3]) def expected_state(request): return produce_expected_state(request.param) fixture_zip("state", ['actual_state', 'expected_state']) def test_with_zipped_fixture(state): assert state.actual_state == state.expected_state @pytest_parametrize_plus_zip("actual, expected", ['actual_state', 'expected_state']) def test_with_direct_parametrization(actual, expected): assert actual == expected ```
1medium
Title: Try using orjson, or adding support for orjson, to the base JSONContent Body: https://github.com/ijl/orjson
1medium
Title: GRADIO_SSR_MODE is not documented Body: ### Describe the bug The environment variable `GRADIO_SSR_MODE` is **not** documented in https://www.gradio.app/guides/environment-variables but referenced on these two lines: https://github.com/gradio-app/gradio/blob/3a6151e9422cc3b02b6e815cb5c72f59bebc60c7/gradio/blocks.py#L2460 and https://github.com/gradio-app/gradio/blob/3a6151e9422cc3b02b6e815cb5c72f59bebc60c7/gradio/routes.py#L1621 ### Have you searched existing issues? 🔎 - [X] I have searched and found no existing issues ### Reproduction ```python import gradio as gr # This is not a code issue. The GRADIO_SSR_MODE will be picked up correctly by the Gradio server but it is not documented. ``` ### Screenshot _No response_ ### Logs _No response_ ### System Info ```shell Gradio Environment Information: ------------------------------ Operating System: Linux gradio version: 5.7.1 gradio_client version: 1.5.0 ------------------------------------------------ gradio dependencies in your environment: aiofiles: 23.2.1 anyio: 4.7.0 audioop-lts is not installed. fastapi: 0.115.6 ffmpy: 0.4.0 gradio-client==1.5.0 is not installed. httpx: 0.28.0 huggingface-hub: 0.26.3 jinja2: 3.1.4 markupsafe: 2.1.5 numpy: 2.1.3 orjson: 3.10.12 packaging: 24.2 pandas: 2.2.3 pillow: 11.0.0 pydantic: 2.10.3 pydub: 0.25.1 python-multipart==0.0.12 is not installed. pyyaml: 6.0.2 ruff: 0.8.2 safehttpx: 0.1.6 semantic-version: 2.10.0 starlette: 0.41.3 tomlkit==0.12.0 is not installed. typer: 0.15.1 typing-extensions: 4.12.2 urllib3: 2.2.3 uvicorn: 0.32.1 authlib; extra == 'oauth' is not installed. itsdangerous; extra == 'oauth' is not installed. gradio_client dependencies in your environment: fsspec: 2024.9.0 httpx: 0.28.0 huggingface-hub: 0.26.3 packaging: 24.2 typing-extensions: 4.12.2 websockets: 12.0 ``` ### Severity I can work around it
0easy
Title: AI Studio import paddlehub报错 Body: 欢迎您反馈PaddleHub使用问题,非常感谢您对PaddleHub的贡献! 在留下您的问题时,辛苦您同步提供如下信息: - 版本、环境信息 1)PaddleHub和PaddlePaddle版本:请提供您的PaddleHub和PaddlePaddle版本号,例如PaddleHub1.4.1,PaddlePaddle1.6.2 2)系统环境:请您描述系统类型,例如Linux/Windows/MacOS/,python版本 - 复现信息:如为报错,请给出复现环境、复现步骤 PaddlePaddle 1.8.4 paddlehub 2.2.0 我在AI Studio中运行项目,只要import paddlehub就出错误 Version mismatch in PaddleHub and PaddlePaddle, you need to upgrade PaddlePaddle to version 2.0.0 or above. 如果卸载重新使用其他方式安装,可以import但是使用任何moudle都找不到 AnimeGAN动漫化模型一键应用(含动漫化小程序体验)这个项目
1medium
Title: Pandas Integer Type Doesn't Convert in Dataset Body: ### What happened? Converted a Pandas `DataFrame` containing a column of type `pandas.Int64Dtype()` into an Xarray `Dataset`. The data variable doesn't get converted to an Xarray compatible type: ``` Data variables: 0 (dim_0) Int64 27B <class 'xarray.core.extension_array.PandasExte... ``` Additionally, this causes an exception if the Dataset is pickled and subsequently loaded: `RecursionError: maximum recursion depth exceeded` ### What did you expect to happen? The data variable ends up as `int64` type. Pickling the Dataset works properly. ### Minimal Complete Verifiable Example ```Python import pandas as pd import xarray as xr import pickle df = pd.DataFrame([1, 2, 3], dtype=pd.Int64Dtype()) ds = xr.Dataset(df) dsdump = pickle.dumps(ds) pickle.loads(dsdump) ``` ### MVCE confirmation - [X] Minimal example — the example is as focused as reasonably possible to demonstrate the underlying issue in xarray. - [X] Complete example — the example is self-contained, including all data and the text of any traceback. - [X] Verifiable example — the example copy & pastes into an IPython prompt or [Binder notebook](https://mybinder.org/v2/gh/pydata/xarray/main?urlpath=lab/tree/doc/examples/blank_template.ipynb), returning the result. - [X] New issue — a search of GitHub Issues suggests this is not a duplicate. - [X] Recent environment — the issue occurs with the latest version of xarray and its dependencies. ### Relevant log output ```Python --------------------------------------------------------------------------- RecursionError Traceback (most recent call last) Cell In[1], line 8 6 ds = xr.Dataset(df) 7 dsdump = pickle.dumps(ds) ----> 8 pickle.loads(dsdump) File ~/metis-dev/.venv/lib/python3.12/site-packages/xarray/core/extension_array.py:112, in PandasExtensionArray.__getattr__(self, attr) 111 def __getattr__(self, attr: str) -> object: --> 112 return getattr(self.array, attr) File ~/metis-dev/.venv/lib/python3.12/site-packages/xarray/core/extension_array.py:112, in PandasExtensionArray.__getattr__(self, attr) 111 def __getattr__(self, attr: str) -> object: --> 112 return getattr(self.array, attr) [... skipping similar frames: PandasExtensionArray.__getattr__ at line 112 (2974 times)] File ~/metis-dev/.venv/lib/python3.12/site-packages/xarray/core/extension_array.py:112, in PandasExtensionArray.__getattr__(self, attr) 111 def __getattr__(self, attr: str) -> object: --> 112 return getattr(self.array, attr) RecursionError: maximum recursion depth exceeded ``` ### Anything else we need to know? Xarray 2024.9.0 does not exhibit this behavior. ### Environment <details> INSTALLED VERSIONS ------------------ commit: None python: 3.12.7 (main, Oct 1 2024, 11:15:50) [GCC 14.2.1 20240910] python-bits: 64 OS: Linux OS-release: 6.6.32-1-lts machine: x86_64 processor: byteorder: little LC_ALL: None LANG: en_US.UTF-8 LOCALE: ('en_US', 'UTF-8') libhdf5: None libnetcdf: None xarray: 2024.10.0 pandas: 2.2.3 numpy: 1.26.4 scipy: 1.14.1 netCDF4: None pydap: None h5netcdf: None h5py: None zarr: 2.18.3 cftime: None nc_time_axis: None iris: None bottleneck: 1.4.2 dask: 2024.10.0 distributed: None matplotlib: 3.9.2 cartopy: None seaborn: None numbagg: None fsspec: 2024.10.0 cupy: None pint: None sparse: None flox: None numpy_groupies: None setuptools: 75.3.0 pip: 23.3.1 conda: None pytest: None mypy: None IPython: 8.29.0 sphinx: None </details>
2hard
Title: Plugin: 用户 Body: ### PyPI 项目名 nonebot-plugin-user ### 插件 import 包名 nonebot_plugin_user ### 标签 [] ### 插件配置项 _No response_
1medium
Title: Option to disable magic surrounding __init__ and __new__? Body: #### Problem Description I'm trying out pdoc for the first time, and was getting frustrated and confused by all the magic handling around the `__init__` and `__new__` methods, so I spent a lot of time playing with the code to figure out how it works just enough to disable it. However, I'm now dependent on a modified local copy of pdoc, which is unfortunate. It would be convenient if it were instead implemented as an official option. I created a mini-project to demonstrate what I mean: https://github.com/odigity/pdocdemo/tree/main/testcases/constructors I created a module [testmodule.py](https://github.com/odigity/pdocdemo/blob/main/testcases/constructors/testmodule.py) with every permutation to test, and added before/after screenshots, which I'll inline at the bottom for convenience. Here's the commit showing what I changed in my local copy of pdoc: https://github.com/odigity/pdocdemo/commit/46567cbc6cc0ba49aac6a524f6acbe37a452298b Specifically: * Stopped pdoc from treating an inherited `__init__` as if it's part of the inheriting class. (`doc.Class._declarations`) * Stopped pdoc from playing with the identity and docstrings of `__new__` vs `__init__`. (`doc.Class._member_objects`) * Stopped pdoc from presenting `__init__` as a constructor named after the class with no `self` in signature. (`function` and `nav_members` macros in module template) * Removed "Inherited Members" section. (`module_contents` block in in module template) In principle, what I want to achieve is for the docs to actually reflect the source. If I define a custom `__init__` in a class, I want it to show up in the docs - otherwise I don't. As for the "Inherited Members" section - I primarily removed it to reduce visual noise, though putting it into a pre-collapsed section that can be manually re-opened might be an option I explore in the future. The other reason is because the `is_public` macro still treats `__init__` as public - I didn't change that because I agree with it - but it means that almost every "Inherited Members" section now shows `__init__`, which is noisy because it adds no value. Sorry for the ramblefest. Hope it made sense. EDIT: I should mention these docs are for internal use (to document our own Django project), not user-facing, like docs for a library would be. That may explain why my preferences are slightly out of alignment with the defaults. ![before](https://user-images.githubusercontent.com/155179/139958655-2639eb8a-daa0-4787-81d1-d7f3f9f4558f.png) ![after](https://user-images.githubusercontent.com/155179/139958650-7d24732d-062c-4237-bf97-3831f6d83713.png)
1medium
Title: Which hyperparameters are suitable for me? Body: ### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and found no similar questions. ### Question Hello. I've already done finetuning and just trained YOLOv11 from scratch. But I have the following problem. Your pretrained model works well with the `car` class in some scenes that I need, but poorly in others that I also need. When I do finetuning of your pretrained model, for some reason, where your model coped well, the quality drops, and where it did not recognize anything, everything is fine. For some reason, somehow finetuning spoils what is already good and improves what was bad. I want to adapt YOLOv11 to work at night. Can you tell me what hyperparameters I need to set so that everything is fine and the way I need it? YOLOv4 just does what it needs to do for some reason. And I want a newer version of YOLO. Maybe I need to freeze something or turn on augmentation? Here is my training startup configuration: ``` task: detect mode: train model: yolov11m.yaml data: ./yolov11_custom.yaml epochs: 500 time: null patience: 100 batch: 32 imgsz: 640 save: true save_period: -1 val_period: 1 cache: false device: 0 workers: 8 project: /YOLOv11_m_night_640 name: yolov11_custom_night exist_ok: false pretrained: true optimizer: auto verbose: true seed: 0 deterministic: true single_cls: false rect: false cos_lr: false close_mosaic: 10 resume: false amp: true fraction: 1.0 profile: false freeze: null multi_scale: false overlap_mask: true mask_ratio: 4 dropout: 0.0 val: true split: val save_json: false save_hybrid: false conf: null iou: 0.7 max_det: 300 half: false dnn: false plots: true source: null vid_stride: 1 stream_buffer: false visualize: false augment: false agnostic_nms: false classes: null retina_masks: false embed: null show: false save_frames: false save_txt: false save_conf: false save_crop: false show_labels: true show_conf: true show_boxes: true line_width: null format: torchscript keras: false optimize: false int8: false dynamic: false simplify: false opset: null workspace: 4 nms: false lr0: 0.01 lrf: 0.01 momentum: 0.937 weight_decay: 0.0005 warmup_epochs: 3.0 warmup_momentum: 0.8 warmup_bias_lr: 0.1 box: 7.5 cls: 0.5 dfl: 1.5 pose: 12.0 kobj: 1.0 label_smoothing: 0.0 nbs: 64 hsv_h: 0.015 hsv_s: 0.7 hsv_v: 0.4 degrees: 0.0 translate: 0.1 scale: 0.5 shear: 0.0 perspective: 0.0 flipud: 0.0 fliplr: 0.5 bgr: 0.0 mosaic: 1.0 mixup: 0.0 copy_paste: 0.0 auto_augment: randaugment erasing: 0.4 crop_fraction: 1.0 cfg: null tracker: botsort.yaml save_dir: /YOLOv11_m_night_640 ``` my `yolov11_custom.yaml`: ``` path: ./data train: ./data/train.txt val: /data/val.txt # Classes names: 0: trailer 1: train 2: trafficlight 3: sign 4: bus 5: truck 6: person 7: bicycle 8: motorcycle 9: car 10: streetlight ``` @glenn-jocher @Y-T-G and others. Please help me. ### Additional _No response_
2hard
Title: here I have error when I run your project in my Mac machine Body: Last Error Received: Process: VR Architecture If this error persists, please contact the developers with the error details. Raw Error Details: ValueError: "zero-size array to reduction operation maximum which has no identity" Traceback Error: " File "UVR.py", line 4716, in process_start File "separate.py", line 686, in seperate File "lib_v5/spec_utils.py", line 112, in normalize File "numpy/core/_methods.py", line 40, in _amax " Error Time Stamp [2023-07-31 16:18:29] Full Application Settings: vr_model: 1_HP-UVR aggression_setting: 10 window_size: 512 batch_size: Default crop_size: 256 is_tta: False is_output_image: False is_post_process: False is_high_end_process: False post_process_threshold: 0.2 vr_voc_inst_secondary_model: No Model Selected vr_other_secondary_model: No Model Selected vr_bass_secondary_model: No Model Selected vr_drums_secondary_model: No Model Selected vr_is_secondary_model_activate: False vr_voc_inst_secondary_model_scale: 0.9 vr_other_secondary_model_scale: 0.7 vr_bass_secondary_model_scale: 0.5 vr_drums_secondary_model_scale: 0.5 demucs_model: Choose Model segment: Default overlap: 0.25 shifts: 2 chunks_demucs: Auto margin_demucs: 44100 is_chunk_demucs: False is_chunk_mdxnet: False is_primary_stem_only_Demucs: False is_secondary_stem_only_Demucs: False is_split_mode: True is_demucs_combine_stems: True demucs_voc_inst_secondary_model: No Model Selected demucs_other_secondary_model: No Model Selected demucs_bass_secondary_model: No Model Selected demucs_drums_secondary_model: No Model Selected demucs_is_secondary_model_activate: False demucs_voc_inst_secondary_model_scale: 0.9 demucs_other_secondary_model_scale: 0.7 demucs_bass_secondary_model_scale: 0.5 demucs_drums_secondary_model_scale: 0.5 demucs_pre_proc_model: No Model Selected is_demucs_pre_proc_model_activate: False is_demucs_pre_proc_model_inst_mix: False mdx_net_model: UVR-MDX-NET Inst HQ 1 chunks: Auto margin: 44100 compensate: Auto is_denoise: False is_invert_spec: False is_mixer_mode: False mdx_batch_size: 5 mdx_voc_inst_secondary_model: No Model Selected mdx_other_secondary_model: No Model Selected mdx_bass_secondary_model: No Model Selected mdx_drums_secondary_model: No Model Selected mdx_is_secondary_model_activate: False mdx_voc_inst_secondary_model_scale: 0.9 mdx_other_secondary_model_scale: 0.7 mdx_bass_secondary_model_scale: 0.5 mdx_drums_secondary_model_scale: 0.5 is_save_all_outputs_ensemble: True is_append_ensemble_name: False chosen_audio_tool: Manual Ensemble choose_algorithm: Min Spec time_stretch_rate: 2.0 pitch_rate: 2.0 is_gpu_conversion: True is_primary_stem_only: False is_secondary_stem_only: False is_testing_audio: False is_add_model_name: False is_accept_any_input: False is_task_complete: False is_normalization: False is_create_model_folder: False mp3_bit_set: 320k save_format: WAV wav_type_set: PCM_16 help_hints_var: False model_sample_mode: True model_sample_mode_duration: 30 demucs_stems: All Stems
2hard
Title: [BUG] No module named 'reco_utils' Body: ### Description Error `No module named 'reco_utils'` in docker test https://github.com/microsoft/recommenders/tree/main/tools/docker ``` $ docker run -it recommenders:cpu pytest tests/unit -m "not spark and not gpu and not notebooks" ImportError while loading conftest '/root/recommenders/tests/conftest.py'. tests/conftest.py:22: in <module> from reco_utils.common.constants import ( E ModuleNotFoundError: No module named 'reco_utils' ``` ### In which platform does it happen? Linux cpu docker
1medium
Title: In Oct 2023 - drop support for k8s 1.23 Body: This is an issue to help us remember the policy (see #2591, #2979, #3040) to allow ourselves to drop support for k8s versions as long as major cloud providers has stopped supporting them. k8s | [GKE EOL](https://endoflife.date/google-kubernetes-engine) | [EKS EOL](https://endoflife.date/amazon-eks) | [AKS EOL](https://docs.microsoft.com/en-gb/azure/aks/supported-kubernetes-versions?tabs=azure-cli#aks-kubernetes-release-calendar) | Comment -|-|-|-|- 1.23 | 31 Jul 2023 | 11 Oct 2023 | Apr 2023 | We can drop support 01 Oct 2023 1.24 | 31 Oct 2023 | 31 Jan 2024 | Jul 2023 | - As part of dropping support for k8s 1.23, we can also adjust this configuration that has a fallback for k8s 1.23 or lower. https://github.com/jupyterhub/zero-to-jupyterhub-k8s/blob/2a02c402f2a23521f9538aea9a934264da3cc959/jupyterhub/templates/scheduling/user-scheduler/configmap.yaml#L9-L23
1medium
Title: SA kernel missing model attribute Body: Trying to load an MCMC run using an SA kernel into `arviz`, I ran into this issue: ``` def test_model(...) ... kernel = SA(test_model) mcmc_test = MCMC(kernel, ...) mcmc_test.run(...) data_test = az.from_numpyro(mcmc_test) ``` which raises... ```--------------------------------------------------------------------------- AttributeError Traceback (most recent call last) /tmp/ipykernel_20194/2118154136.py in <module> ----> 1 data_test = az.from_numpyro(mcmc_test) ~/miniconda3/envs/refit_fvs/lib/python3.9/site-packages/arviz/data/io_numpyro.py in from_numpyro(posterior, prior, posterior_predictive, predictions, constant_data, predictions_constant_data, coords, dims, pred_dims, num_chains) 331 Number of chains used for sampling. Ignored if posterior is present. 332 """ --> 333 return NumPyroConverter( 334 posterior=posterior, 335 prior=prior, ~/miniconda3/envs/refit_fvs/lib/python3.9/site-packages/arviz/data/io_numpyro.py in __init__(self, posterior, prior, posterior_predictive, predictions, constant_data, predictions_constant_data, coords, dims, pred_dims, num_chains) 91 self._samples = samples 92 self.nchains, self.ndraws = posterior.num_chains, posterior.num_samples ---> 93 self.model = self.posterior.sampler.model 94 # model arguments and keyword arguments 95 self._args = self.posterior._args # pylint: disable=protected-access AttributeError: 'SA' object has no attribute 'model' ``` Looking at the source code and noticing that this works for NUTS and HMC, the missing piece in the SA class seems to be the `model` property... going to prepare a PR to add it in the same manner as it's found in NUTS and HMC classes.
1medium
Title: [FEATURE] Upgrade nni and scikit-learn Body: ### Description <!--- Describe your expected feature in detail --> - Upgrade nni from current version (1.5) so that we can use scikit-learn>=0.22.1 - Add nni back to the core option of the pypi package
1medium
Title: Why is first so slow for default SQLAlchemyConnectionField Body: ``` requests_monitor = SQLAlchemyConnectionField( RequestConnection, request_name=String(), ... ) def resolve_requests_monitor(self, info, **args): query = RequestNode.get_query(info) query = query.join(Request.dataset) for field, value in args.items(): query = query.filter(getattr(Request, field) == value) return query ``` The table I was querying has 10 million records. When I ran the following query, it was super slow. But when I added `query.limit(1000)`, it was way faster. This made me wonder if first and last were not applied by default when querying the db and actually applied afterwards. If not, what's the correct way of using it? ``` requestsMonitor(first:1000){ edges{ node{ requestId } } } ```
1medium
Title: Additional open AI configuation options for editor AI chat and code completion Body: ### Description Adding gitlab issue as per discord conversation at https://discord.com/channels/1059888774789730424/1349354519284289536 Setup config `.toml` to add additional optional keywords for httpx session that the openai python library uses under the hood for openAI compatible API/AI assistants that are secured with ssl/tls. Specifically, add in `verify` and `ssl_verify` as keywords for openai in config. See https://www.python-httpx.org/advanced/ssl/ for more examples/options. ### Suggested solution Based on the discord conversation, it looks like that the changes could be made in: `marimo/_server/api/endpoints/ai.py` `marimo/_config/config.py` I am unsure if the openai client creation can directly be based these parameters or if an httpx_client object would have to be passed to the `http_client` parameter when calling ``` OpenAI( default_headers={"api-key": key}, api_key=key, base_url=base_url, ) ``` ### Alternative _No response_ ### Additional context I am uncertain if I will be able to make the pull request, due to some corporate rules we have to follow when contributing (formally at least) to open source projects.
1medium
Title: Image Mask is not working since update Body: ### Describe the bug When using stroke in the image mask to mask an area, it stops and not making any mask. this is a issue since gradio 5.12.0 ### Have you searched existing issues? 🔎 - [x] I have searched and found no existing issues ### Reproduction ```python import gradio as gr ``` ### Screenshot _No response_ ### Logs ```shell ``` ### System Info ```shell Gradio Environment Information: ------------------------------ Operating System: Linux gradio version: 5.11.0 gradio_client version: 1.5.3 ------------------------------------------------ gradio dependencies in your environment: aiofiles: 23.2.1 anyio: 4.6.0 audioop-lts is not installed. fastapi: 0.115.5 ffmpy: 0.4.0 gradio-client==1.5.3 is not installed. httpx: 0.27.2 huggingface-hub: 0.28.1 jinja2: 3.1.4 markupsafe: 2.1.5 numpy: 1.26.4 orjson: 3.10.7 packaging: 24.2 pandas: 2.2.3 pillow: 10.4.0 pydantic: 2.9.2 pydub: 0.25.1 python-multipart: 0.0.19 pyyaml: 6.0.2 ruff: 0.9.6 safehttpx: 0.1.6 semantic-version: 2.10.0 starlette: 0.41.3 tomlkit: 0.12.0 typer: 0.12.5 typing-extensions: 4.12.2 urllib3: 2.2.3 uvicorn: 0.30.6 authlib; extra == 'oauth' is not installed. itsdangerous; extra == 'oauth' is not installed. gradio_client dependencies in your environment: fsspec: 2024.9.0 httpx: 0.27.2 huggingface-hub: 0.28.1 packaging: 24.2 typing-extensions: 4.12.2 websockets: 12.0 ``` ### Severity I can work around it
1medium
Title: LightningCLI, WandbLogger and save_hyperparameters() inconsistencies Body: ### Bug description There are issues when using WandbLogger from a LightningCLI configuration. First, full configuration is not automatically forwarded to the kwargs of the WandbLogger, so configuration is not automatically saved. This can be fixed with a custom CLI: ```python class CustomSaveConfigCallback(SaveConfigCallback): # Saves full training configuration def save_config( self, trainer: Trainer, pl_module: LightningModule, stage: _LITERAL_WARN ) -> None: for logger in trainer.loggers: if issubclass(type(logger), WandbLogger): logger.experiment.config.update(self.config.as_dict()) return super().save_config(trainer, pl_module, stage) ``` However, there will still be duplicate hyperparameters on wandb: parameters saved with `save_hyperparameters` are not nested within the relative model or dataclass, but are placed in the root (because `save_hyperparameters` feeds the logger a flattened list). `save_hyperparameters` should place the updated parameters in the correct config path on wandb, instead of duplicating them at the root. Logging can be disabled by subclassing WandbLogger ```python class CustomWandbLogger(WandbLogger): # Disable unintended hyperparameter logging (already saved on init) def log_hyperparams(self, *args, **kwargs): ... ``` but then updated (or additional) hyperparameters added when initializing the model won't be stored. Maybe there's already a better way to fix this behavior (or is it indended)? ### What version are you seeing the problem on? v2.2 ### How to reproduce the bug Config file: ```python trainer: logger: class_path: WandbLogger ``` Inside a model e.g. ``` class ConvolutionalNetwork(L.LightningModule): def __init__( self, dim_in: int, num_internal_channels: int, num_layers: int, kernel_size: int, num_neurons_dense: int, seq_len: int, ): super().__init__() self.save_hyperparameters() ```
1medium
Title: Default to first cell or prefix a new cell for parameters Body: When a user doesn't specify a parameter tag cell it would be nice if papermill defaulted to some sane logical setting. Nominally parameterizing the beginning of the notebook seems reasonable and would make adoption of existing notebooks quicker when they have naive inputs.
1medium
Title: Unable to load transformer.wmt18.en-de Body: ## 🐛 Bug Loading transformer.wmt18.en-de` from the hub fails despite working in #1287. The checkpoint has to be specified manually (otherwise a different error appears because `model.pt` does not exist in the archive) but that errors as well. ### To Reproduce ```python3 import torch torch.hub.load( 'pytorch/fairseq', 'transformer.wmt18.en-de', checkpoint_file='wmt18.model1.pt', tokenizer='moses', bpe='subword_nmt' ) ``` Resulting error output: ``` Traceback (most recent call last): File "/home/vilda/.local/lib/python3.10/site-packages/hydra/_internal/config_loader_impl.py", line 513, in _apply_overrides_to_config OmegaConf.update(cfg, key, value, merge=True) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/omegaconf.py", line 613, in update root.__setattr__(last_key, value) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/dictconfig.py", line 285, in __setattr__ raise e File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/dictconfig.py", line 282, in __setattr__ self.__set_impl(key, value) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/dictconfig.py", line 266, in __set_impl self._set_item_impl(key, value) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/basecontainer.py", line 398, in _set_item_impl self._validate_set(key, value) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/dictconfig.py", line 143, in _validate_set self._validate_set_merge_impl(key, value, is_assign=True) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/dictconfig.py", line 156, in _validate_set_merge_impl self._format_and_raise( File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/base.py", line 95, in _format_and_raise format_and_raise( File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/_utils.py", line 694, in format_and_raise _raise(ex, cause) File "/home/vilda/.local/lib/python3.10/site-packages/omegaconf/_utils.py", line 610, in _raise raise ex # set end OC_CAUSE=1 for full backtrace omegaconf.errors.ValidationError: child 'checkpoint.save_interval_updates' is not Optional full_key: checkpoint.save_interval_updates reference_type=CheckpointConfig object_type=CheckpointConfig The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/home/vilda/mt-metric-estimation/./src/retranslate.py", line 100, in <module> model = MODELS[args.model](args.direction) File "/home/vilda/mt-metric-estimation/./src/retranslate.py", line 88, in <lambda> "w18t": lambda direction: FairSeqWrap(config=f"transformer.wmt18.{direction}"), File "/home/vilda/mt-metric-estimation/./src/retranslate.py", line 61, in __init__ torch.hub.load( File "/home/vilda/.local/lib/python3.10/site-packages/torch/hub.py", line 540, in load model = _load_local(repo_or_dir, model, *args, **kwargs) File "/home/vilda/.local/lib/python3.10/site-packages/torch/hub.py", line 569, in _load_local model = entry(*args, **kwargs) File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/models/fairseq_model.py", line 267, in from_pretrained x = hub_utils.from_pretrained( File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/hub_utils.py", line 82, in from_pretrained models, args, task = checkpoint_utils.load_model_ensemble_and_task( File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/checkpoint_utils.py", line 425, in load_model_ensemble_and_task state = load_checkpoint_to_cpu(filename, arg_overrides) File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/checkpoint_utils.py", line 343, in load_checkpoint_to_cpu state = _upgrade_state_dict(state) File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/checkpoint_utils.py", line 681, in _upgrade_state_dict state["cfg"] = convert_namespace_to_omegaconf(state["args"]) File "/home/vilda/.local/lib/python3.10/site-packages/fairseq/dataclass/utils.py", line 399, in convert_namespace_to_omegaconf composed_cfg = compose("config", overrides=overrides, strict=False) File "/home/vilda/.local/lib/python3.10/site-packages/hydra/experimental/compose.py", line 31, in compose cfg = gh.hydra.compose_config( File "/home/vilda/.local/lib/python3.10/site-packages/hydra/_internal/hydra.py", line 507, in compose_config cfg = self.config_loader.load_configuration( File "/home/vilda/.local/lib/python3.10/site-packages/hydra/_internal/config_loader_impl.py", line 151, in load_configuration return self._load_configuration( File "/home/vilda/.local/lib/python3.10/site-packages/hydra/_internal/config_loader_impl.py", line 277, in _load_configuration ConfigLoaderImpl._apply_overrides_to_config(config_overrides, cfg) File "/home/vilda/.local/lib/python3.10/site-packages/hydra/_internal/config_loader_impl.py", line 520, in _apply_overrides_to_config raise ConfigCompositionException( hydra.errors.ConfigCompositionException: Error merging override checkpoint.save_interval_updates=null ``` ### Environment - fairseq Version: 0.12.2 - PyTorch Version: 1.12.1+cu116 - OS: Ubuntu Linux 22.04 - How you installed fairseq: pip - Python version: 3.10.4 - CUDA/cuDNN version: 11.6 (note that this issue is unrelated to #4680)
2hard
Title: Type annotation `def f(a: List[int])` not working in codon.jit decorated function Body: The code ``` import codon @codon.jit def f(a: List[int]): return len(a) ``` raises a `NameError: name 'List' is not defined`, while the same function definition works fine in a pure codon program. Also, other type annotations like `int` work in codon.jit decorated functions. Do I misunderstand something, or is this a bug or missing feature in v0.16.1?
1medium
Title: CI test linux://doc:doc_code_cgraph_profiling is consistently_failing Body: CI test **linux://doc:doc_code_cgraph_profiling** is consistently_failing. Recent failures: - https://buildkite.com/ray-project/postmerge/builds/8711#01956a36-520a-4bea-a351-c4b739770614 - https://buildkite.com/ray-project/postmerge/builds/8711#01956a0e-a7de-4ac8-a743-147f7e001fb9 DataCaseName-linux://doc:doc_code_cgraph_profiling-END Managed by OSS Test Policy
1medium
Title: Change features_only to feature_defs_only Body: When looking at the [DFS](https://featuretools.alteryx.com/en/stable/generated/featuretools.dfs.html#featuretools-dfs) call, I feel that the `features_only` option is misleading. Setting this to `True` only returns definitions and not the feature matrix. So I believe the option should be `feature_defs_only` #### Code Example ```python import featuretools as ft es = ft.demo.load_mock_customer(return_entityset=True) feature_defs = ft.dfs( entityset=es, target_dataframe_name="customers", agg_primitives=["mean"], trans_primitives=["time_since_previous"], features_defs_only=True, ) feature_defs ```
0easy
Title: Only display observed categories in legend Body: Hi, This is a feature request **Current behaviour** When plotting with a `pd.CategoricalDtype(categories=categories)` column as `hue` or `style`, all the elements in `categories` are displayed in the legend, even the categories that don't appear in the data. **Desired feature** It would be nice to also have the possibility to only display the observed categories in the legend. For example with a new parameter in `FacetGrid.add_legend`: ```python g.add_legend(observed_only=True) ``` **Relevance** This would be useful in case you have multiple plots involving the same categories but not all categories observed in each plot. This way, the colours/style would be coherent between the plots, without overcrowding the legends.
1medium
Title: The drop table operation does not trigger the before_drop and after_drop events Body: **Describe the bug** The `table.drop` and `op.drop_table` operation does not trigger the `before_drop` and `after_drop` events. **To Reproduce** Create function as follow: ``` @event.listens_for(Table, "before_drop") def record_before_event(target, conn, **kwargs): print("Will never be printed!") ``` drop table `using table.drop()` or `op.drop_table('table')` (if migrations). The function won't be invoked. **Expected behavior** Table drop table should trigger an event.
1medium
Title: Remove ultralytics dependency Body: ### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/orgs/ultralytics/discussions) and found no similar questions. ### Question I am using python 3.6.9 version and i need to load best.pt of yolo11n model , how to remove the dependency of ultralytics model and load best.pt model ### Additional _No response_
1medium
Title: Trade Updates Body: Question on this part of the algorithm: ```python if event == 'fill': if data.order['side'] == 'buy': position.update_total_shares( int(data.order['filled_qty']) ) else: position.update_total_shares( -1 * int(data.order['filled_qty']) ) position.remove_pending_order( data.order['id'], data.order['side'] ) ``` I'm wondering if it should really be: ```python if event == 'fill': if data.order['side'] == 'buy': position.update_filled_amount( data.order['id'], int(data.order['filled_qty']), data.order['side'] ) else: position.update_filled_amount( data.order['id'], int(data.order['filled_qty']), data.order['side'] ) position.remove_pending_order( data.order['id'], data.order['side'] ) ``` Because, I believe, a PARTIALLY_FILLED order can then become FILL order once it is fully filled. Any thoughts or do I have that incorrect?
1medium
Title: Support bybit candles -> is closed Body: **Is your feature request related to a problem? Please describe.** NO **Describe the solution you'd like** I would like to store only confirmed data from bybit. **Describe alternatives you've considered** Add following to bybit.py def __init__(self, candle_closed_only=False, **kwargs): super().__init__(**kwargs) self.candle_closed_only = candle_closed_only self.ws_defaults['compression'] = None Edit following in bybit.py async def _candle(self, msg: dict, timestamp: float): ''' { "topic": "klineV2.1.BTCUSD", //topic name "data": [{ "start": 1572425640, //start time of the candle "end": 1572425700, //end time of the candle "open": 9200, //open price "close": 9202.5, //close price "high": 9202.5, //max price "low": 9196, //min price "volume": 81790, //volume "turnover": 8.889247899999999, //turnover "confirm": False, //snapshot flag "cross_seq": 297503466, "timestamp": 1572425676958323 //cross time }], "timestamp_e6": 1572425677047994 //server time } ''' symbol = self.exchange_symbol_to_std_symbol(msg['topic'].split(".")[-1]) ts = msg['timestamp_e6'] / 1_000_000 for entry in msg['data']: > if self.candle_closed_only and entry['confirm']: c = Candle(self.id, symbol, entry['start'], entry['end'], self.candle_interval, None, Decimal(entry['open']), Decimal(entry['close']), Decimal(entry['high']), Decimal(entry['low']), Decimal(entry['volume']), > entry['confirm'], ts, raw=entry) await self.callback(CANDLES, c, timestamp) Can you please add this. Thank you and best regards.
1medium
Title: [Trio] Task groups are marked inactive early Body: Code: ```py import anyio async def second(): print("This never happens") async def first(tg): await anyio.sleep(0) await tg.spawn(second) async def main(): async with anyio.create_task_group() as tg: await tg.spawn(first, tg) anyio.run(main, backend="trio") ``` Error: ```py Traceback (most recent call last): File "/home/laura/dev/discord/curious/aaa.py", line 18, in <module> anyio.run(main, backend="trio") File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/anyio/__init__.py", line 112, in run return asynclib.run(func, *args, **backend_options) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/trio/_core/_run.py", line 1337, in run raise runner.main_task_outcome.error File "/home/laura/dev/discord/curious/aaa.py", line 15, in main await tg.spawn(first, tg) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/async_generator/_util.py", line 42, in __aexit__ await self._agen.asend(None) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/async_generator/_impl.py", line 366, in step return await ANextIter(self._it, start_fn, *args) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/async_generator/_impl.py", line 202, in send return self._invoke(self._it.send, value) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/async_generator/_impl.py", line 209, in _invoke result = fn(*args) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/anyio/_backends/trio.py", line 90, in create_task_group tg._active = False File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/trio/_core/_run.py", line 397, in __aexit__ raise combined_error_from_nursery File "/home/laura/dev/discord/curious/aaa.py", line 10, in first await tg.spawn(second) File "/home/laura/.local/share/virtualenvs/curious-yiBjxk2Q/lib/python3.7/site-packages/anyio/_backends/trio.py", line 78, in spawn raise RuntimeError('This task group is not active; no new tasks can be spawned.') RuntimeError: This task group is not active; no new tasks can be spawned. ``` Both curio and asyncio backends allow this to work.
1medium
Title: Weibull distribution Body: I got ValueError: Expected parameter scale (Tensor of shape (32, 96, 1)) of distribution Weibull(scale: torch.Size([32, 96, 1]), concentration: torch.Size([32, 96, 1])) to satisfy the constraint GreaterThan(lower_bound=0.0), but found invalid values:tensor([[[ 3.9661e-30],[ 1.2568e+01],[ 6.9787e-08]........while using Weibull likelihood parameter but the input was given while using Laplace and Gaussian Likelihood & code was run successfully.
2hard
Title: [Question] Seed variable in analyze function Body: Hi, I am using the sobol analyze (https://salib.readthedocs.io/en/latest/_modules/SALib/analyze/sobol.html#analyze) I would like to know, what is the purpose of variable seed in the sobol analyze? Thank you!
3misc
Title: TFLearn model not predicting when using gunicorn Body: I have the following class as part of my project: class LSTMSentimentClassifier(SentimentClassifier): def __init__(self, model_name: str, model_location: str, wordvec_model_location: str, lexicon_location: str): super(LSTMSentimentClassifier, self).__init__("LSTM_{}".format(model_name)) self.logger = logging.getLogger(__name__) self.model_name = model_name self.model_location = model_location self.numDimensions = 300 self.num_hidden = 128 self.learning_rate = 0.001 self.drop_out = 0.8 tf.reset_default_graph() net = input_data(shape=[None, self.numDimensions]) net = embedding(net, input_dim=20000, output_dim=self.num_hidden) net = tflearn.lstm(net, self.num_hidden, return_seq=True) net = dropout(net, self.drop_out) net = tflearn.lstm(net, self.num_hidden, return_seq=True) net = dropout(net, self.drop_out) net = fully_connected(net, 2, activation='softmax') net = regression(net, optimizer='adam', learning_rate=self.learning_rate, loss='mean_square') self.nn = tflearn.DNN(net, clip_gradients=0., tensorboard_verbose=0) self.nn.load(model_location) self.wordvec_model = KeyedVectors.load_word2vec_format(wordvec_model_location) self.lexicon = pickle.load(open(lexicon_location, 'rb')) self.logger.info("Inside NRCSentimentClassifier -- " "Loaded NRC model file for the model: {}".format(model_name)) def word_to_vec(self, sentence: str) -> np.array: if type(sentence) == str: sentence = sentence else: sentence = sentence[0] words = sentence.split() featureVec = [0] * self.numDimensions for word in words: try: try: featureVec = featureVec + (self.wordvec_model[word] + self.lexicon[word]) except: featureVec = featureVec + self.wordvec_model[word] except: featureVec Vec = [max(0, val) for val in featureVec] return np.array(Vec) def predict_sentiment(self, text: str) -> list: text_vec = self.word_to_vec(text) sentiment_classes = ["positive", "negative"] labels_with_probabilities = {} model_probabilities = self.nn.predict([text_vec])[0] for label, probability in zip(sentiment_classes, model_probabilities): labels_with_probabilities[label] = probability labels_with_probabilities_sorted = sorted(labels_with_probabilities.items(), key=lambda x: x[1], reverse=True) print(labels_with_probabilities_sorted) output = [] for label, probability in labels_with_probabilities_sorted: label_val = str(label) probability_val = float(probability) output.append(SentimentScore(sentence=text, sentiment_class=SentimentClass.get_label_NRC(label_val), confidence_score=probability_val, sentiment_strength=1.0)) return output When I run this class as part of a python script eg `python classifier.py` it works with a main method and calling the class with the file paths, giving predictions for sentences that I pass into it, `model.predict_sentiment(sentence)`, where `model = LSTMSentimentClassifier("lstm", path_to_classifier, path_to_wordembeddings, path_to_lexicon)` However this classifier is part of a larger project that uses `gunicorn` and `falcon`. I can start the project and it loads fine, but when I make an api call using the classifier, the classifier gets stuck at the predict. I know that the model has loaded correctly but I am confused as to why I can use the above class when running the single python file, but the class predict doesn't work when using it as part of the gunicorn process. The file locations are correct. I have used prints to verify that a sentence is coming in, it is correctly being turned into a word vector in the correct form and passed into the `nn.predict`, however it is here that the class stops working and gunicorn times out. Any help would be much appreciated.
2hard
Title: [BUG] - Exception: Invalid Response Body: I am trying to use with Django, it was working last week, but now it is showing an error with ``` Exception: Invalid Response ```. I upgrade the API too. via ``` pip install TikTokApi --upgrade ```. but still showing exception. Here's the code ``` API = TikTokApi() results = API.trending(count=VIDEOS_COUNT) ``` **Error Trace** ``` {'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.111 Safari/537.36', 'accept-encoding': 'gzip, deflate, br', 'accept': 'application/json, text/plain, */*', 'Connection': 'keep-alive', 'authority': 'm.tiktok.com', 'method': 'GET', 'path': '/api/item_list/?aid=1988&app_name=tiktok_web&device_platform=web&referer=https%3A%2F%2Fwww.', 'scheme': 'https', 'accept-language': 'en-US,en;q=0.9', 'referer': 'https://www.tiktok.com/', 'sec-fetch-dest': 'empty', 'sec-fetch-mode': 'cors', 'sec-fetch-site': 'same-site', 'cookie': 'tt_webid_v2=746436216'} Converting response to JSON failed response is below (probably empty) illegal request... Internal Server Error Traceback (most recent call last): File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 103, in getData return r.json() File "/home/{user}/Documents/venv/lib/python3.6/site-packages/requests/models.py", line 900, in json return complexjson.loads(self.text, **kwargs) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 190, in trending res = self.getData(b, **kwargs) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 113, in getData raise Exception("Invalid Response") Exception: Invalid Response ERROR 2020-12-03 00:01:27,081 log 16905 140297062823680 Internal Server Error: /dashboard/ Traceback (most recent call last): File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 103, in getData return r.json() File "/home/{user}/Documents/venv/lib/python3.6/site-packages/requests/models.py", line 900, in json return complexjson.loads(self.text, **kwargs) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/core/handlers/exception.py", line 34, in inner response = get_response(request) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/core/handlers/base.py", line 115, in _get_response response = self.process_exception_by_middleware(e, request) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/core/handlers/base.py", line 113, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/lib/python3.6/contextlib.py", line 52, in inner return func(*args, **kwds) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/views/generic/base.py", line 71, in view return self.dispatch(request, *args, **kwargs) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/views/generic/base.py", line 97, in dispatch return handler(request, *args, **kwargs) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/django/views/generic/edit.py", line 142, in post return self.form_valid(form) File "/home/{user}/{DjangoApp}/{DjangoApp}/dashboard/views.py", line 25, in form_valid 'hashtag': hashtag, File "/home/{user}/{DjangoApp}/{DjangoApp}/utils/tiktok.py", line 86, in generate_video results = API.trending(count=VIDEOS_COUNT) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 190, in trending res = self.getData(b, **kwargs) File "/home/{user}/Documents/venv/lib/python3.6/site-packages/TikTokApi/tiktok.py", line 113, in getData raise Exception("Invalid Response") Exception: Invalid Response ``` **Desktop (please complete the following information):** - OS: Ubuntu 18.04 - TikTokApi Version latest **Additional context**
2hard
Title: Warning in EDA Body: During the analysis of `robert` dataset from OpenML: ```py Using a non-tuple sequence for multidimensional indexing is deprecated; use `arr[tuple(seq)]` instead of `arr[seq]`. In the future this will be interpreted as an array index, `arr[np.array(seq)]`, which will result either in an error or a different result. ```
0easy
Title: Manaphy evolution trigger Body: There is no evolution trigger for Manaphy, evolution-chain/250 and pokemon/490 (It should be, being grown up with a ditto with him)
1medium
Title: Lost remote browser access Body: Not sure what happened but cannot access mycodo through the browser anymore and get this error message: Version: 8.12.9 Database: 0187ea22dc4b Model: Raspberry Pi 3 Model B Rev 1.2 Release: Distributor ID: Raspbian Description: Raspbian GNU/Linux 10 (buster) Release: 10 Codename: buster Firmware: Dec 1 2021 15:08:00 Copyright (c) 2012 Broadcom version 71bd3109023a0c8575585ba87cbb374d2eeb038f (clean) (release) (start_x) Error (Full Traceback): Traceback (most recent call last): File "/var/mycodo-root/env/lib/python3.7/site-packages/flask/app.py", line 2073, in wsgi_app response = self.full_dispatch_request() File "/var/mycodo-root/env/lib/python3.7/site-packages/flask/app.py", line 1519, in full_dispatch_request return self.finalize_request(rv) File "/var/mycodo-root/env/lib/python3.7/site-packages/flask/app.py", line 1540, in finalize_request response = self.process_response(response) File "/var/mycodo-root/env/lib/python3.7/site-packages/flask/app.py", line 1888, in process_response self.session_interface.save_session(self, ctx.session, response) File "/var/mycodo-root/env/lib/python3.7/site-packages/flask_session/sessions.py", line 367, in save_session total_seconds(app.permanent_session_lifetime)) File "/var/mycodo-root/env/lib/python3.7/site-packages/cachelib/file.py", line 147, in set self._prune() File "/var/mycodo-root/env/lib/python3.7/site-packages/cachelib/file.py", line 87, in _prune if self._threshold == 0 or not self._file_count > self._threshold: File "/var/mycodo-root/env/lib/python3.7/site-packages/cachelib/file.py", line 58, in _file_count return self.get(self._fs_count_file) or 0 File "/var/mycodo-root/env/lib/python3.7/site-packages/cachelib/file.py", line 127, in get return pickle.load(f) EOFError: Ran out of input I am running off master so kot sure if this has something to do with it
2hard
Title: Support "ls" command and endpoint definitions Body: I've been using this program for a few days now and overall I love it. I'm using it to test endpoints while developing my API. One issue I have however is that whenever I type `cd`, my muscle memory kicks in and I immediately follow with an `ls` out of habit. It's always a little disappointing when this doesn't work :( I was thinking of ways to solve this, and it would be great if I could pass in my endpoint definition file ala ``` bash $ http-prompt --specification=my_api_spec.json ``` This would enable http-prompt to do a bunch of smart things like suggest endpoints when I type `ls` or autocomplete parameter names. Could also add a `help` command to view the documentation for any given endpoint. There are a couple of widely used spec formats, with Swagger probably being the most popular http://nordicapis.com/top-specification-formats-for-rest-apis/
1medium
Title: `ImportError: Module "drf_yasg.generators" does not define a "OpenAPISchemaGenerator" attribute/class` after upgrading DRF==3.14.0 Body: # Bug Report ## Description After update Django Rest Framework to 3.14.0, Django did not start, because drf-yasg raise exception `ImportError: Could not import 'drf_yasg.generators.OpenAPISchemaGenerator' for API setting 'DEFAULT_GENERATOR_CLASS'. ` Reverting to drf==3.13.1 resolves the issue. ## Is this a regression? Not sure, there was a similar issue here: https://github.com/axnsan12/drf-yasg/issues/641 ## Minimal Reproduction ```python # settings.py INSTALLED_APPS = [ ... "drf_yasg", ] SWAGGER_SETTINGS = { "SECURITY_DEFINITIONS": { "Bearer": { "type": "apiKey", "in": "header", "name": "Authorization", "template": "Bearer {apiKey}", }, }, "DEFAULT_FIELD_INSPECTORS": [ "drf_yasg.inspectors.CamelCaseJSONFilter", "drf_yasg.inspectors.InlineSerializerInspector", "drf_yasg.inspectors.RelatedFieldInspector", "drf_yasg.inspectors.ChoiceFieldInspector", "drf_yasg.inspectors.FileFieldInspector", "drf_yasg.inspectors.DictFieldInspector", "drf_yasg.inspectors.SimpleFieldInspector", "drf_yasg.inspectors.StringDefaultFieldInspector", ], } ``` ```python # urls.py from drf_yasg import openapi from drf_yasg.views import get_schema_view from rest_framework import permissions schema_view = get_schema_view( openapi.Info( title="My API", default_version="v1", description="My API", terms_of_service="https://www.google.com/policies/terms/", contact=openapi.Contact(email="leohakim@gmail.com"), license=openapi.License(name="BSD License"), ), public=True, permission_classes=(permissions.AllowAny,), ) ``` ## Stack trace / Error message ```code File "/usr/local/lib/python3.9/site-packages/drf_yasg/views.py", line 67, in get_schema_view _generator_class = generator_class or swagger_settings.DEFAULT_GENERATOR_CLASS File "/usr/local/lib/python3.9/site-packages/drf_yasg/app_settings.py", line 122, in __getattr__ val = perform_import(val, attr) File "/usr/local/lib/python3.9/site-packages/rest_framework/settings.py", line 166, in perform_import return import_from_string(val, setting_name) File "/usr/local/lib/python3.9/site-packages/rest_framework/settings.py", line 180, in import_from_string raise ImportError(msg) ImportError: Could not import 'drf_yasg.generators.OpenAPISchemaGenerator' for API setting 'DEFAULT_GENERATOR_CLASS'. ImportError: Module "drf_yasg.generators" does not define a "OpenAPISchemaGenerator" attribute/class. ``` ## Your Environment ```code drf-yasg=1.21.3 djangorestframework=3.14.0 django=4.1.1 ```
2hard
Title: DataFrame's column names display Unicode Body: # DataFrame's column names display Unicode I am using JupyterLab. I encountered an encoding problem when importing a data set. The column names of the DataFrame contain Chinese, but the Chinese is displayed as Unicode. The strange thing is that the value of the field also contains Chinese, but it can be displayed normally. <img width="772" alt="DXTRyHFcaW" src="https://user-images.githubusercontent.com/59652488/221740595-5ae81295-b90c-43bd-a747-9eecceeef7db.png">
1medium
Title: There are two instances of different models. If one instance executes db.session.commit (), the other instance immediately print (obj .__ dict__) will be empty. Body: --- ### Expected behavior model .__ dict__ Fields where data can be obtained ```sh root@c57f84919524:/webroot# python manage.py shell >>> from app.models import * >>> u = User.query.filter().first() >>> img = ProjectImages.query.filter().first() >>> print(img.__dict__) {'_sa_instance_state': <sqlalchemy.orm.state.InstanceState object at 0x7f5b40df7040>, 'id': 1, 'created_at': datetime.datetime(2020, 2, 20, 21, 34, 14), 'image': 'https://cdn.yttys.com/item_1520582553119_0649002690.jpg', 'project_id': 1, 'delete_flag': 0, 'sort': 1} >>> from app import db >>> db.session.add(u) >>> db.session.commit() >>> print(img.__dict__) {'_sa_instance_state': <sqlalchemy.orm.state.InstanceState object at 0x7f5b40df7040>} >>> ``` ### Actual behavior ```python obj = model.query.filter().first() other = model2.query.filter().first() db.session.add(other) db.session.commit() print(obj .__ dict__) ``` The final output is: ``` dict_keys (['_ sa_instance_state']) ``` There are two instances of different models. If one instance executes db.session.commit (), the other instance immediately print (obj .__ dict__) will be empty. ### surroundings * operating system: * Python version: 3.8.1 * Flask-SQLAlchemy version: 2.4.1 * SQLAlchemy version: 1.3.13
2hard
Title: dash_table.DataTable. When fixed_rows={'headers': True, 'data': 0}, the table gets distorted Body: When fixed_rows is set to headers = True the page does not resize and is by default distorted (see image on the right with False value compared with the image on the left with True). I use latest version of dash, dash 1.16.2 ![Screenshot from 2020-09-28 22-28-15](https://user-images.githubusercontent.com/26002644/94477162-2f1f2280-01da-11eb-8cc8-820b15106e3d.png) `app.layout = html.Div([ dash_table.DataTable( id='datatable-interactivity', columns=[ {"name": i, "id": i, "deletable": True, "selectable": True} for i in df.columns ], data=df.to_dict('records'), editable=False, hidden_columns=False, filter_action="native", sort_action="native", sort_mode="multi", column_selectable="multi", row_selectable=False, row_deletable=False, selected_columns=[], selected_rows=[], page_action="native", page_current=0, fixed_rows={'headers': True, 'data': 0},`
1medium
Title: training using own data [ Body: <!-- Please respect the title [Discussion] tag. --> Hi, if i want to use another my own dataset of songs instead of musdb18, whats the steps i should do in using spleeter?
1medium
Title: Insert / Update Hive Query failing. Body: I am facing this issue with pyHive. I have connected to Hive on EMR as follows ```from pyhive import hive config = {'hive.txn.manager':'org.apache.hadoop.hive.ql.lockmgr.DbTxnManager', 'hive.support.concurrency': 'true', 'hive.enforce.bucketing': 'true', 'hive.exec.dynamic.partition.mode': 'nonstrict', 'hive.compactor.initiator.on': 'true',} connection = hive.connect ("X.X.X.X", configuration= config) cursor = connection.cursor() cursor.execute("Update table set x=2 where y=1") ``` I get the following exception Stack trace and I'm not able to point out what the problem is. ``` TExecuteStatementResp(status=TStatus(statusCode=3, infoMessages=[ '*org.apache.hive.service.cli.HiveSQLException:Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.tez.TezTask:28:27', 'org.apache.hive.service.cli.operation.Operation:toSQLException:Operation.java:380', 'org.apache.hive.service.cli.operation.SQLOperation:runQuery:SQLOperation.java:257', 'org.apache.hive.service.cli.operation.SQLOperation:runInternal:SQLOperation.java:293', 'org.apache.hive.service.cli.operation.Operation:run:Operation.java:320', 'org.apache.hive.service.cli.session.HiveSessionImpl:executeStatementInternal:HiveSessionImpl.java:530', 'org.apache.hive.service.cli.session.HiveSessionImpl:executeStatement:HiveSessionImpl.java:506', 'sun.reflect.GeneratedMethodAccessor16:invoke::-1', 'sun.reflect.DelegatingMethodAccessorImpl:invoke:DelegatingMethodAccessorImpl.java:43', 'java.lang.reflect.Method:invoke:Method.java:498', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:78', 'org.apache.hive.service.cli.session.HiveSessionProxy:access$000:HiveSessionProxy.java:36', 'org.apache.hive.service.cli.session.HiveSessionProxy$1:run:HiveSessionProxy.java:63', 'java.security.AccessController:doPrivileged:AccessController.java:-2', 'javax.security.auth.Subject:doAs:Subject.java:422', 'org.apache.hadoop.security.UserGroupInformation:doAs:UserGroupInformation.java:1926', 'org.apache.hive.service.cli.session.HiveSessionProxy:invoke:HiveSessionProxy.java:59', 'com.sun.proxy.$Proxy35:executeStatement::-1', 'org.apache.hive.service.cli.CLIService:executeStatement:CLIService.java:280', 'org.apache.hive.service.cli.thrift.ThriftCLIService:ExecuteStatement:ThriftCLIService.java:531', 'org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement:getResult:TCLIService.java:1437', 'org.apache.hive.service.rpc.thrift.TCLIService$Processor$ExecuteStatement:getResult:TCLIService.java:1422', 'org.apache.thrift.ProcessFunction:process:ProcessFunction.java:39', 'org.apache.thrift.TBaseProcessor:process:TBaseProcessor.java:39', 'org.apache.hive.service.auth.TSetIpAddressProcessor:process:TSetIpAddressProcessor.java:56', 'org.apache.thrift.server.TThreadPoolServer$WorkerProcess:run:TThreadPoolServer.java:286', 'java.util.concurrent.ThreadPoolExecutor:runWorker:ThreadPoolExecutor.java:1149', 'java.util.concurrent.ThreadPoolExecutor$Worker:run:ThreadPoolExecutor.java:624', 'java.lang.Thread:run:Thread.java:748'], sqlState='08S01', errorCode=1, errorMessage='Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.tez.TezTask'), operationHandle=None) ``` When I run the **exact same query on EMR it is executing successfully** (The table has been set up as an ACID table stored as ORC and transactions enabled) Select queries are working fine with the connection set up as shown above. I seem to get the same error on inserts and updates. I'm not sure if there is a configuration that I'm missing to add to my connection or if it is an issue on Pyhive. Any help would be appreciated. Thanks.
2hard
Title: [BUG] `use_arrow_dtype` doesn't work when read a directory path for `read_parquet` Body: <!-- Thank you for your contribution! Please review https://github.com/mars-project/mars/blob/master/CONTRIBUTING.rst before opening an issue. --> **Describe the bug** `use_arrow_dtype` doesn't work when read a directory path for `read_parquet` **To Reproduce** ``` Python In [26]: df = md.read_parquet(tempdir, use_arrow_dtype=True) In [27]: df.dtypes Out[27]: a int64 b object c float64 dtype: object ``` Expected dtypes should be: ``` Python In [14]: df.dtypes Out[14]: a int64 b Arrow[string] c float64 dtype: object ``` **Expected behavior** A clear and concise description of what you expected to happen. **Additional context** Add any other context about the problem here.
1medium
Title: Where is 28x28 mask to binary mask conversion Body: I am a beginner in Mask-RCNN. The model predicts a 28x28 segmentation mask then converted to a binary mask of coordinates of the original image to superimpose on it. I cannot understand what code does that job? Kindly guide.
1medium
Title: Addon error: Flow is not killable Body: #### Problem Description ~~~ > mitmproxy -r sky-phone-login.txt Errors logged during startup: Addon error: Flow is not killable. Traceback (most recent call last): File "mitmproxy\addons\disable_h2c.py", line 41, in request File "mitmproxy\addons\disable_h2c.py", line 33, in process_flow File "mitmproxy\flow.py", line 235, in kill mitmproxy.exceptions.ControlException: Flow is not killable. ~~~ #### Steps to reproduce the behavior: ~~~ : save.file @shown sky-phone-login.txt ~~~ #### System Information ~~~ Mitmproxy: 10.3.0 binary Python: 3.12.3 OpenSSL: OpenSSL 3.2.1 30 Jan 2024 Platform: Windows-10-10.0.18363-SP0 ~~~
1medium
Title: AttributeError: 'str' object has no attribute 'decode' Body: When I ran 'saving.load_weights_from_hdf5_group_by_name(f, layers)',which is the line 2135 of mrcnn/model.py,the AttributeError raised as: 'str' object has no attribute 'decode'
1medium
Title: Memory accumulation on the GPU - and a Fix! Body: I experienced OOM with very big datasets and realised the images were accumulating on the GPU. After a lot of digging I found that here: https://github.com/nerfstudio-project/nerfstudio/blob/702886b213a5942b64250b3c4dc632e62237a511/nerfstudio/data/datamanagers/full_images_datamanager.py#L347 Because a deepcopy of the original tensor is not made, the cpu_cached images are transferred to the GPU and left there taking up memory quickly. Modifying the data dictionary does modify the underlying cached_train. Interestingly a deepcopy is performed here: https://github.com/nerfstudio-project/nerfstudio/blob/702886b213a5942b64250b3c4dc632e62237a511/nerfstudio/data/datamanagers/full_images_datamanager.py#L373 I would argue that we don't want the dataset to accumulate on the GPU and a deepcopy should be added.
1medium
Title: basculin is not found Body: - Provide a clear description of the issue. There is no entry for basculin - pokemon number 550. - Provide a clear description of the steps to reproduce. 1. Search for basculin https://pokeapi.co/api/v2/pokemon/basculin 2. Returns undefined - Provide a clear description of the expected behavior. I expected info on basculin.
1medium
Title: Is there any request quota limit per IP to httpbin.org/status ? Body: I am writing a small js/html5 app to monitor my internet connection. For this I need to "ping" an external service to check if the device is connected to the internet. Httpbin/status/200 seems to be just what I was looking for, however my app will ping the service every 3 seconds or so (e.g. that means a peak of 28k times a day). For me only using it, I guess that will be okay, however I plan to release it on github, so hundreds of users will probably start to use it. Do you have any quota limit in place that I should be aware of before releasing it?
1medium
Title: html.Script not rendering the javascript code Body: Hi, I'm trying to run a javascript code wrapped in `html.Script`. But it's not rendering the JS code. ``` recharts_js = """ const { BarChart, Bar, XAxis, YAxis, Tooltip, Legend, CartesianGrid, ResponsiveContainer } = Recharts; const data = [ { dmu: "dmu1", "Efficiency score": 100, Status: "Efficient" }, { dmu: "dmu2", "Efficiency score": 78, Status: "InEfficient" }, { dmu: "dmu3", "Efficiency score": 100, Status: "Efficient" }, { dmu: "dmu4", "Efficiency score": 100, Status: "Efficient" }, { dmu: "dmu5", "Efficiency score": 89, Status: "InEfficient" }, { dmu: "dmu6", "Efficiency score": 95, Status: "InEfficient" }, ]; class CustomBarChart extends React.Component { render() { return ( <Recharts.BarChart width={600} height={400} data={data} margin={{ top: 20, right: 30, left: 20, bottom: 5 }} > <Recharts.CartesianGrid strokeDasharray="3 3" /> <Recharts.XAxis dataKey="dmu" /> <Recharts.YAxis /> <Recharts.Tooltip /> <Recharts.Legend payload={[ { value: "Efficient", type: "square", id: "ID01", color: "#FFA500" }, // Orange for Efficient { value: "InEfficient", type: "square", id: "ID02", color: "#32CD32" }, // Green for InEfficient ]} /> <Recharts.Bar dataKey="Efficiency score"> {data.map((entry, index) => ( <Recharts.Cell key={`cell-${index}`} fill={entry.Status === "Efficient" ? "#FFA500" : "#32CD32"} /> ))} </Recharts.Bar> </Recharts.BarChart> ); } } ReactDOM.render(<CustomBarChart />, document.getElementById('recharts-container')); """ html.Div( [ html.Script(children=recharts_js), ], ) ```
2hard
Title: Change in the Neighbours widget Body: @BlazZupan started a discussion about the Neighbours widgets. He basically, wanted an output Data with distances to *all* data instance, and overlooked that he can achieve the same effect by unchecking the checkbox that limits the number of neighbours. Here are some notes: **Neighbours**: - Add missing distances (normalized Euclidean and Manhatta). (Note to self: an input with distances is not a solution because it would incur quadratic time complexity). - Add an output Data with all instance and their distances to the reference, and **make it lazy**. The output is superfluous, but somehow consistent with other widgets that output subsets, and it's harmless if lazy. - Perhaps: remove the checkbox, keep line edit. This is still a matter of discussion; removing the checkbox requires some code for backward compatibilitty. We also discussed how to provide distances to multiple references. One option was an additional input to Distances of type Data Table that would represent reference examples), and an output with distances to all those instance. The problems with this solution are - potentially many meta columns with distances - Distances can currently be connected only to a few widgets, which is nice. With the additional output, it can be connected to almost everything, so the pop-up is populated with lots and lots of widgets. - We want additional options, like a column that contains an id of the closest reference, or some kind of aggregation (min, max, avg) of distances to refrences. We haven't decided anything, but it seems this calls for a new widget.
1medium
Title: uint16 SegmentationMapOnImage Body: I think it is quite common to encode masks using uint16, therefore I would suggest to enable this format for the `SegmentationMapOnImage` class. It seems like this format is not supported yet: https://github.com/aleju/imgaug/blob/69ac72ef4f2b9d5de62c7813dcf3427ca4a604b5/imgaug/imgaug.py#L4814
1medium
Title: Support sqlite database urls relative to project path. Body: When developing it makes sense to have the database file in the local repository directory, to have a `dev.env` which is valid for all developers it needs to be relative.. I have been doing hacks in settings.py to get around it for a few years now but I would like a discussion about how to maybe integrate such a feature into django-environ. This basically just assumes that if the the file is at the root of the system it's instead relative but I'm looking for a way to get rid of these in every project.. ```py for db in DATABASES: if DATABASES[db]['ENGINE'] == 'django.db.backends.sqlite3': if os.path.split(DATABASES[db]['NAME'])[0] == "/": DATABASES[db]['NAME'] = root(DATABASES[db]['NAME'][1:]) ``` I am thinking that this is kind of sane because no one proably stores the database in the root of the file system but making it optional by having a `sqlite_root` argument to the db-url function maybe works? ```py root = environ.Path(__file__) - 2 ... DATABASES = { 'default': env.db(sqlite_root=root()), } ``` I think that both relative AND absolute paths should work. One option which I think might work safely for triggering a relative path instead of absolute is to do it when host is `.` or some other reserved name which isn't really a hostname. It should error out if `sqlite_root` was not provided to the url parse function. ```py DATABASE_URL=sqlite://./db.sqlite3 ``` Please share your opinions on this..
1medium
Title: [DOCS] Encounter ValueError When Following User Guide Annotation Label Schema Body: ### URL(s) with the issue [User Guide: annotation label schema](https://docs.voxel51.com/user_guide/annotation.html#label-schema) ### Description of proposal (what needs changing) **Description of Error** I encountered following error when I use the schema provided in user guide `ValueError: Attribute 'attr3' of type 'text' requires a list of values` **Potential Fix** I resolved the error by adding in `"values":[]` after line 24. But I'm not sure if this is the best approach to address this problem. The code example would be look like: ``` anno_key = "..." label_schema = { "new_field": { "type": "classifications", "classes": ["class1", "class2"], "attributes": { "attr1": { "type": "select", "values": ["val1", "val2"], "default": "val1", }, "attr2": { "type": "radio", "values": [True, False], "default": False, } }, }, "existing_field": { "classes": ["class3", "class4"], "attributes": { "attr3": { "type": "text", "values":[] } } }, } dataset.annotate(anno_key, label_schema=label_schema) ``` **Other questions** I'm using label schema to construct annotation task in CVAT. My understanding is that by including `existing_field`, I could modify existing annotation on CVAT. However, when I do so I got: `A field with label type 'detections' is already being annotated. Ignoring field 'ground_truth'...` (ground_truth here is my `existing_field`) and the `existing_field` is not loaded to CVAT. Can someone enlighten me where I made a mistake or direct me to relevant docs. Thanks in advance ### Willingness to contribute The FiftyOne Community encourages documentation contributions. Would you or another member of your organization be willing to contribute a fix for this documentation issue to the FiftyOne codebase? - [ ] Yes. I can contribute a documentation fix independently - [x] Yes. I would be willing to contribute a documentation fix with guidance from the FiftyOne community - [ ] No. I cannot contribute a documentation fix at this time
1medium
Title: Getting 403 while connecting to websocket Body: I have used python-socketio with FastAPI & Uvicorn as server, But when I try to connect to socket using the below mentioned URL getting 403 as response Dependency: python-engineio 3.14.2 python-socketio 4.6.1 uvicorn 0.22.0 fastapi 0.97.0 URL to connect: ws://localhost:5050/ws **Code:** ``` import uvicorn from engineio import ASGIApp from fastapi import FastAPI from socketio import AsyncServer app = FastAPI(title=__name__) sio = AsyncServer(async_mode="asgi", cors_allowed_origins="*", always_connect=True) socket_app = ASGIApp(sio, other_asgi_app=app) app.mount(path='/ws', app=socket_app, name="SocketServer") if __name__ == "__main__": uvicorn.run( "__main__:app", host="0.0.0.0", port=5050, reload=False ) ``` ***Response*** Error: Unexpected server response: 403
1medium
Title: Network and config issues when starting jupyterhub Body: I have a working jupyterhub installation (based mostly on Andrea Zonca's instructions [https://zonca.github.io/2016/04/jupyterhub-sdsc-cloud.html] and it worked flawlessly for a machine-learning workshop. However, I now want to upgrade to using jupyterlab with my hub, and unfortunately experienced much the same problems as https://github.com/jupyterhub/jupyterhub-deploy-docker/issues/26. I could access jupyterlab from the hub by manually changing the url once I had logged in, but was unable to have it as the default url. Hence I'm trying your docker deployment of the hub :) My first issue arises with nginx on the host. If I have the service running and I start the hub I get the following error: ``` ERROR: for jupyterhub Cannot start service hub: driver failed programming external connectivity on endpoint jupyterhub (dd0d984e99ddb8ce10900a489c82f96d42ab9f84e9b4977d874c85f0e0374096): Error starting userland proxy: listen tcp 0.0.0.0:443: bind: address already in use ERROR: for hub Cannot start service hub: driver failed programming external connectivity on endpoint jupyterhub (dd0d984e99ddb8ce10900a489c82f96d42ab9f84e9b4977d874c85f0e0374096): Error starting userland proxy: listen tcp 0.0.0.0:443: bind: address already in use ERROR: Encountered errors while bringing up the project. ``` If I stop nginx, the hub starts with no errors, but the site won't load. The logs do however show an error with the config: ``` docker logs jupyterhub [E 2017-09-11 02:17:41.976 JupyterHub application:569] Exception while loading config file /srv/jupyterhub/jupyterhub_config.py Traceback (most recent call last): File "/opt/conda/lib/python3.5/site-packages/traitlets/config/application.py", line 557, in _load_config_files config = loader.load_config() File "/opt/conda/lib/python3.5/site-packages/traitlets/config/loader.py", line 457, in load_config self._read_file_as_dict() File "/opt/conda/lib/python3.5/site-packages/traitlets/config/loader.py", line 489, in _read_file_as_dict py3compat.execfile(conf_filename, namespace) File "/opt/conda/lib/python3.5/site-packages/ipython_genutils/py3compat.py", line 185, in execfile exec(compiler(f.read(), fname, 'exec'), glob, loc) File "/srv/jupyterhub/jupyterhub_config.py", line 80, in <module> name = parts[0] IndexError: list index out of range [I 2017-09-11 02:17:41.980 JupyterHub app:745] Writing cookie_secret to /srv/jupyterhub/jupyterhub_cookie_secret [W 2017-09-11 02:17:42.302 JupyterHub app:365] Generating CONFIGPROXY_AUTH_TOKEN. Restarting the Hub will require restarting the proxy. Set CONFIGPROXY_AUTH_TOKEN env or JupyterHub.proxy_auth_token config to avoid this message. [W 2017-09-11 02:17:42.336 JupyterHub app:864] No admin users, admin interface will be unavailable. [W 2017-09-11 02:17:42.337 JupyterHub app:865] Add any administrative users to `c.Authenticator.admin_users` in config. [I 2017-09-11 02:17:42.337 JupyterHub app:892] Not using whitelist. Any authenticated user will be allowed. [I 2017-09-11 02:17:42.360 JupyterHub app:1453] Hub API listening on http://127.0.0.1:8081/hub/ [W 2017-09-11 02:17:42.364 JupyterHub app:1174] Running JupyterHub without SSL. I hope there is SSL termination happening somewhere else... [I 2017-09-11 02:17:42.364 JupyterHub app:1176] Starting proxy @ http://*:8000/ 02:17:42.628 - info: [ConfigProxy] Proxying http://*:8000 to http://127.0.0.1:8081 02:17:42.633 - info: [ConfigProxy] Proxy API at http://127.0.0.1:8001/api/routes [I 2017-09-11 02:17:42.675 JupyterHub app:1485] JupyterHub is now running at http://127.0.0.1:8000/ ``` So, to me it looks as though the userlist isn't there? When I build the image it says it's copying it, so not sure whats happening. I'm doing all this on a VM running on Openstack with Ubuntu 16.04, and I'm using Docker version 17.06.1-ce, build 874a737, docker-compose version 1.16.1, build 6d1ac21. I want to use PAM and local user accounts so I commented out references to oauth in the Dockerfile and Makefile (previously I was using the systemuser rather than the singleuser, is this another potential source of problems?). I plan on trying kubernetes and helm in the future, but having never used either, this seemed like the easiest and quickest way to get what I want. Any advice on how to proceed greatly appreciated.
2hard
Title: create table + reflection for parenthesis around server defaults that contain spaces, all DBs Body: ### Describe the bug When using `metadata.reflect` on a sqlite3 table with a "_timestamp_" column defined as: `timestamp DATETIME NOT NULL DEFAULT (datetime(CURRENT_TIMESTAMP, 'localtime'))` the resulting metadata is: `datetime(CURRENT_TIMESTAMP, 'localtime')` without the surrounding parentheses and `metadata.create_all(engine)` will fail. ### Optional link from https://docs.sqlalchemy.org which documents the behavior that is expected _No response_ ### SQLAlchemy Version in Use 2.0.38. ### DBAPI (i.e. the database driver) pysqlite ### Database Vendor and Major Version SQLite ### Python Version Python 3.12.3 ### Operating system Linux ### To Reproduce ```python import sqlite3 conn: sqlite3.Connection = sqlite3.connect("tmp.db") # Create a table conn.execute("""CREATE TABLE IF NOT EXISTS recipients (key INTEGER PRIMARY KEY, "campaign_id" INTEGER, "campaign_name" TEXT, "email_id" TEXT, "send_date" TEXT, "delivered_date" TEXT, "open_date" TEXT, "total_opens" INTEGER, "total_apple_mpp_opens" INTEGER, "unsubscribe_date" TEXT, "hard_bounce_date" TEXT, "hard_bounce_reason" TEXT, "soft_bounce_date" TEXT, "soft_bounce_reason" TEXT, "open_ip" TEXT, "click_ip" TEXT, "unsubscribe_ip" TEXT, "clicked_links_count" INTEGER, "complaint_date" TEXT, "click_date" TEXT, "click_url" TEXT, timestamp DATETIME NOT NULL DEFAULT (datetime(CURRENT_TIMESTAMP, 'localtime')) );""") conn.close() from sqlacodegen.generators import SQLModelGenerator from sqlalchemy import MetaData, create_engine # Define the SQLite URL sqlite_url = "sqlite:///tmp.db" engine = create_engine(sqlite_url) # Reflect metadata metadata = MetaData() # Reflect db metadata.reflect(bind=engine) new_engine = create_engine("sqlite:///tmp2.db") metadata.create_all(new_engine ``` ### Error ``` # Copy the complete stack trace and error message here, including SQL log output if applicable. ``` --------------------------------------------------------------------------- OperationalError Traceback (most recent call last) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1964, in Connection._exec_single_context(self, dialect, context, statement, parameters) 1963 if not evt_handled: -> 1964 self.dialect.do_execute( 1965 cursor, str_statement, effective_parameters, context 1966 ) 1968 if self._has_events or self.engine._has_events: File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/default.py:942, in DefaultDialect.do_execute(self, cursor, statement, parameters, context) 941 def do_execute(self, cursor, statement, parameters, context=None): --> 942 cursor.execute(statement, parameters) OperationalError: near "(": syntax error The above exception was the direct cause of the following exception: OperationalError Traceback (most recent call last) Cell In[34], line 2 1 new_engine = create_engine("sqlite:///tmp2.db") ----> 2 metadata.create_all(new_engine) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/schema.py:5907, in MetaData.create_all(self, bind, tables, checkfirst) 5883 def create_all( 5884 self, 5885 bind: _CreateDropBind, 5886 tables: Optional[_typing_Sequence[Table]] = None, 5887 checkfirst: bool = True, 5888 ) -> None: 5889 """Create all tables stored in this metadata. 5890 5891 Conditional by default, will not attempt to recreate tables already (...) 5905 5906 """ -> 5907 bind._run_ddl_visitor( 5908 ddl.SchemaGenerator, self, checkfirst=checkfirst, tables=tables 5909 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:3249, in Engine._run_ddl_visitor(self, visitorcallable, element, **kwargs) 3242 def _run_ddl_visitor( 3243 self, 3244 visitorcallable: Type[Union[SchemaGenerator, SchemaDropper]], 3245 element: SchemaItem, 3246 **kwargs: Any, 3247 ) -> None: 3248 with self.begin() as conn: -> 3249 conn._run_ddl_visitor(visitorcallable, element, **kwargs) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:2456, in Connection._run_ddl_visitor(self, visitorcallable, element, **kwargs) 2444 def _run_ddl_visitor( 2445 self, 2446 visitorcallable: Type[Union[SchemaGenerator, SchemaDropper]], 2447 element: SchemaItem, 2448 **kwargs: Any, 2449 ) -> None: 2450 """run a DDL visitor. 2451 2452 This method is only here so that the MockConnection can change the 2453 options given to the visitor so that "checkfirst" is skipped. 2454 2455 """ -> 2456 visitorcallable(self.dialect, self, **kwargs).traverse_single(element) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/visitors.py:664, in ExternalTraversal.traverse_single(self, obj, **kw) 662 meth = getattr(v, "visit_%s" % obj.__visit_name__, None) 663 if meth: --> 664 return meth(obj, **kw) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/ddl.py:928, in SchemaGenerator.visit_metadata(self, metadata) 926 for table, fkcs in collection: 927 if table is not None: --> 928 self.traverse_single( 929 table, 930 create_ok=True, 931 include_foreign_key_constraints=fkcs, 932 _is_metadata_operation=True, 933 ) 934 else: 935 for fkc in fkcs: File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/visitors.py:664, in ExternalTraversal.traverse_single(self, obj, **kw) 662 meth = getattr(v, "visit_%s" % obj.__visit_name__, None) 663 if meth: --> 664 return meth(obj, **kw) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/ddl.py:966, in SchemaGenerator.visit_table(self, table, create_ok, include_foreign_key_constraints, _is_metadata_operation) 957 if not self.dialect.supports_alter: 958 # e.g., don't omit any foreign key constraints 959 include_foreign_key_constraints = None 961 CreateTable( 962 table, 963 include_foreign_key_constraints=( 964 include_foreign_key_constraints 965 ), --> 966 )._invoke_with(self.connection) 968 if hasattr(table, "indexes"): 969 for index in table.indexes: File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/ddl.py:314, in ExecutableDDLElement._invoke_with(self, bind) 312 def _invoke_with(self, bind): 313 if self._should_execute(self.target, bind): --> 314 return bind.execute(self) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1416, in Connection.execute(self, statement, parameters, execution_options) 1414 raise exc.ObjectNotExecutableError(statement) from err 1415 else: -> 1416 return meth( 1417 self, 1418 distilled_parameters, 1419 execution_options or NO_OPTIONS, 1420 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/sql/ddl.py:180, in ExecutableDDLElement._execute_on_connection(self, connection, distilled_params, execution_options) 177 def _execute_on_connection( 178 self, connection, distilled_params, execution_options 179 ): --> 180 return connection._execute_ddl( 181 self, distilled_params, execution_options 182 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1527, in Connection._execute_ddl(self, ddl, distilled_parameters, execution_options) 1522 dialect = self.dialect 1524 compiled = ddl.compile( 1525 dialect=dialect, schema_translate_map=schema_translate_map 1526 ) -> 1527 ret = self._execute_context( 1528 dialect, 1529 dialect.execution_ctx_cls._init_ddl, 1530 compiled, 1531 None, 1532 exec_opts, 1533 compiled, 1534 ) 1535 if self._has_events or self.engine._has_events: 1536 self.dispatch.after_execute( 1537 self, 1538 ddl, (...) 1542 ret, 1543 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1843, in Connection._execute_context(self, dialect, constructor, statement, parameters, execution_options, *args, **kw) 1841 return self._exec_insertmany_context(dialect, context) 1842 else: -> 1843 return self._exec_single_context( 1844 dialect, context, statement, parameters 1845 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1983, in Connection._exec_single_context(self, dialect, context, statement, parameters) 1980 result = context._setup_result_proxy() 1982 except BaseException as e: -> 1983 self._handle_dbapi_exception( 1984 e, str_statement, effective_parameters, cursor, context 1985 ) 1987 return result File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:2352, in Connection._handle_dbapi_exception(self, e, statement, parameters, cursor, context, is_sub_exec) 2350 elif should_wrap: 2351 assert sqlalchemy_exception is not None -> 2352 raise sqlalchemy_exception.with_traceback(exc_info[2]) from e 2353 else: 2354 assert exc_info[1] is not None File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/base.py:1964, in Connection._exec_single_context(self, dialect, context, statement, parameters) 1962 break 1963 if not evt_handled: -> 1964 self.dialect.do_execute( 1965 cursor, str_statement, effective_parameters, context 1966 ) 1968 if self._has_events or self.engine._has_events: 1969 self.dispatch.after_cursor_execute( 1970 self, 1971 cursor, (...) 1975 context.executemany, 1976 ) File ~/Projects/Imported/Fadim_ltd/2025/zcrm/.venv/lib/python3.12/site-packages/sqlalchemy/engine/default.py:942, in DefaultDialect.do_execute(self, cursor, statement, parameters, context) 941 def do_execute(self, cursor, statement, parameters, context=None): --> 942 cursor.execute(statement, parameters) OperationalError: (sqlite3.OperationalError) near "(": syntax error [SQL: CREATE TABLE campaigns ( "key" INTEGER, id INTEGER, name TEXT, type TEXT, status TEXT, "testSent" INTEGER, header TEXT, footer TEXT, sender TEXT, "replyTo" TEXT, "toField" TEXT, "previewText" TEXT, tag TEXT, "inlineImageActivation" INTEGER, "mirrorActive" INTEGER, recipients TEXT, statistics TEXT, "htmlContent" TEXT, subject TEXT, "scheduledAt" TEXT, "createdAt" TEXT, "modifiedAt" TEXT, "shareLink" TEXT, "sentDate" TEXT, "sendAtBestTime" INTEGER, "abTesting" INTEGER, timestamp DATETIME DEFAULT datetime(CURRENT_TIMESTAMP, 'localtime') NOT NULL, PRIMARY KEY ("key") ) ] (Background on this error at: https://sqlalche.me/e/20/e3q8) ### Additional context _No response_
1medium
Title: loftq_utils.py depdends on huggingface_hub.errors, which doesn't appear in some versions of huggingface_hub Body: ### System Info loftq_utils.py refers to huggingface_hub.errors Should the requirements.txt include huggingface_hub? I have huggingface_hub version 0.19.4 and it does not have huggingface_hub.errors. Is there a workaround or another version of huggingface_hub? ### Who can help? _No response_ ### Information - [X] The official example scripts - [ ] My own modified scripts ### Tasks - [ ] An officially supported task in the `examples` folder - [X] My own task or dataset (give details below) ### Reproduction Should the requirements.txt include huggingface_hub? I have huggingface_hub version 0.19.4 and it does not have huggingface_hub.errors. Is there a workaround or another version of huggingface_hub? ### Expected behavior Should the requirements.txt include huggingface_hub? I have huggingface_hub version 0.19.4 and it does not have huggingface_hub.errors. Is there a workaround or another version of huggingface_hub?
1medium
Title: ChatterBot Body: Can I know what all are the advantages and disadvantages of using ChatterBot to build a bot?
3misc
Title: cache in nfs error Body: ### Describe the bug - When reading dataset, a cache will be generated to the ~/. cache/huggingface/datasets directory - When using .map and .filter operations, runtime cache will be generated to the /tmp/hf_datasets-* directory - The default is to use the path of tempfile.tempdir - If I modify this path to the NFS disk, an error will be reported, but the program will continue to run - https://github.com/huggingface/datasets/blob/main/src/datasets/config.py#L257 ``` Traceback (most recent call last): File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/process.py", line 315, in _bootstrap self.run() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/process.py", line 108, in run self._target(*self._args, **self._kwargs) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/managers.py", line 616, in _run_server server.serve_forever() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/managers.py", line 182, in serve_forever sys.exit(0) SystemExit: 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 300, in _run_finalizers finalizer() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 224, in __call__ res = self._callback(*self._args, **self._kwargs) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 133, in _remove_temp_dir rmtree(tempdir) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 718, in rmtree _rmtree_safe_fd(fd, path, onerror) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 675, in _rmtree_safe_fd onerror(os.unlink, fullname, sys.exc_info()) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 673, in _rmtree_safe_fd os.unlink(entry.name, dir_fd=topfd) OSError: [Errno 16] Device or resource busy: '.nfs000000038330a012000030b4' Traceback (most recent call last): File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/process.py", line 315, in _bootstrap self.run() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/process.py", line 108, in run self._target(*self._args, **self._kwargs) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/managers.py", line 616, in _run_server server.serve_forever() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/managers.py", line 182, in serve_forever sys.exit(0) SystemExit: 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 300, in _run_finalizers finalizer() File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 224, in __call__ res = self._callback(*self._args, **self._kwargs) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/site-packages/multiprocess/util.py", line 133, in _remove_temp_dir rmtree(tempdir) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 718, in rmtree _rmtree_safe_fd(fd, path, onerror) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 675, in _rmtree_safe_fd onerror(os.unlink, fullname, sys.exc_info()) File "/home/wzp/miniconda3/envs/dask/lib/python3.8/shutil.py", line 673, in _rmtree_safe_fd os.unlink(entry.name, dir_fd=topfd) OSError: [Errno 16] Device or resource busy: '.nfs0000000400064d4a000030e5' ``` ### Steps to reproduce the bug ``` import os import time import tempfile from datasets import load_dataset def add_column(sample): # print(type(sample)) # time.sleep(0.1) sample['__ds__stats__'] = {'data': 123} return sample def filt_column(sample): # print(type(sample)) if len(sample['content']) > 10: return True else: return False if __name__ == '__main__': input_dir = '/mnt/temp/CN/small' # some json dataset dataset = load_dataset('json', data_dir=input_dir) temp_dir = '/media/release/release/temp/temp' # a nfs folder os.makedirs(temp_dir, exist_ok=True) # change huggingface-datasets runtime cache in nfs(default in /tmp) tempfile.tempdir = temp_dir aa = dataset.map(add_column, num_proc=64) aa = aa.filter(filt_column, num_proc=64) print(aa) ``` ### Expected behavior no error occur ### Environment info datasets==2.18.0 ubuntu 20.04
2hard
Title: Patch To Embedding correct? Body: In line 95 of ViT [[self.patch_to_embedding = nn.Linear(patch_dim, dim)]] Is it supposed to be a nn.Linear layer? I believe its a learnable tensor. The paper says "E is a trainable linear projection that maps each vectorized patch to the model dimension D". Yannic also referred E as a linear projection matrix. Could you please share your thoughts? NB: I have modified and run E as a nn.parameter type tensor and produces similar results
1medium
Title: Multiple authentication/authorizations Body: Hi there, wondering how does one auhenticate and authorize a route. I'm doing this way: ``` @router.get('auth/', auth=[AuthBearer(), account_required], response=UserSchema) def user_detail(request): return request.user ``` Where `AuthBearer` is a class that extends `HttpBearer` and has an `authenticate` method that checks for the token received (as explained in the docs: https://django-ninja.rest-framework.com/tutorial/authentication/#http-bearer). And `account_required` is a method that checks if user has an account (`hasattr(request.user, 'account') and request.user.account is not None`) (docs: https://django-ninja.rest-framework.com/tutorial/authentication/#custom-function) When I hit this endpoint, code is touched on `AuthBearer.authenticate` but does not on `account_required`. So I assume that, for multiple auth, I must have done something wrong. Since for single auth (`auth=account_required`), the `account_required` method is touched. - If I change from `list` to `set` I got the same result. - It always touches only the one in the first position. If I change to `auth=[account_required, AuthBearer()]`, then `account_required` is touched instead of `AuthBearer.authenticate`. What am I doing wrong? Thank you, and congrats for this amazing library!
1medium
Title: Fix Travis / tests Body: https://travis-ci.org/jakubroztocil/httpie/jobs/385658131#L598-L616
1medium
Title: ci maintenance: queue workflows or jobs with new GitHub workflow `concurrency` option? Body: I think this GitHub workflow thingy can be really useful. It makes two workflows being triggered not run in parallell if they are within the same group. This will make us queue workflows in sequence. ```yaml concurrency: group: ${{ github.head_ref }} ``` This can also be applied to specific jobs, but not steps within jobs. This makes it even more valuable if we manage to split apart our build workflow into multiple separate jobs and then merges them into a single later. By doing so, we can mark the job that does the push to github and could run into merge conflicts marked to belong to a specific concurrency group to ensure that never happens rather than marking the entire workflow with all steps that could happen cuncurrently.
1medium
Title: Distributed LDA Connection refused - Dispatcher host is always 127.0.0.1 Body: ## Problem description I am trying to run the distributed LDA on multiple machines. Following this tutorial: https://radimrehurek.com/gensim/models/lda_worker.html Setup is working but once I run the distributed lda I get the following error: ![error](https://user-images.githubusercontent.com/963798/71465376-c0c9b180-27bc-11ea-9bab-ce061f33a424.PNG) Dispatcher worker recognizes that there are lda-workers available but they cannot connect to the dispatcher. ![gensim](https://user-images.githubusercontent.com/963798/71465420-e8207e80-27bc-11ea-83df-17a063fcbe89.PNG) The dispatcher host is always 127.0.0.1 I have also tried setting the host ip on the dispatcher but no luck ``` python -m gensim.models.lda_dispatcher --host HOST_IP & ``` Nameserver is recognized on other machines and shows the correct ip.
2hard
Title: searching metod Body: ### Actions before raising this issue - [X] I searched the existing issues and did not find anything similar. - [X] I read/searched [the docs](https://docs.cvat.ai/docs/) ### Is your feature request related to a problem? Please describe. https://docs.cvat.ai/docs/manual/advanced/search/ add featuring in textbox search by image name. ### Describe the solution you'd like in the text box you can write the name of image and the filter show the image with the tags directly ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1medium
Title: Dependabot can't resolve your Python dependency files Body: Dependabot can't resolve your Python dependency files. As a result, Dependabot couldn't update your dependencies. The error Dependabot encountered was: ``` ERROR: ERROR: Could not find a version that matches mplfinance>=0.12 Skipped pre-versions: 0.11.1a0, 0.11.1a0, 0.11.1a1, 0.11.1a1, 0.12.0a0, 0.12.0a0, 0.12.0a1, 0.12.0a1, 0.12.0a2, 0.12.0a2, 0.12.0a3, 0.12.0a3, 0.12.3a0, 0.12.3a0, 0.12.3a1, 0.12.3a1, 0.12.3a2, 0.12.3a2, 0.12.3a3, 0.12.3a3, 0.12.4a0, 0.12.4a0, 0.12.5a1, 0.12.5a1, 0.12.5a2, 0.12.5a2, 0.12.5a3, 0.12.5a3, 0.12.6a0, 0.12.6a0, 0.12.6a1, 0.12.6a1, 0.12.6a2, 0.12.6a2, 0.12.6a3, 0.12.6a3, 0.12.7a0, 0.12.7a0, 0.12.7a3, 0.12.7a3, 0.12.7a4, 0.12.7a4, 0.12.7a5, 0.12.7a5, 0.12.7a7, 0.12.7a7, 0.12.7a10, 0.12.7a10, 0.12.7a12, 0.12.7a12, 0.12.7a17, 0.12.7a17 There are incompatible versions in the resolved dependencies. [pipenv.exceptions.ResolutionFailure]: return resolve_deps( [pipenv.exceptions.ResolutionFailure]: File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 718, in resolve_deps [pipenv.exceptions.ResolutionFailure]: resolved_tree, hashes, markers_lookup, resolver = actually_resolve_deps( [pipenv.exceptions.ResolutionFailure]: File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 480, in actually_resolve_deps [pipenv.exceptions.ResolutionFailure]: resolved_tree = resolver.resolve() [pipenv.exceptions.ResolutionFailure]: File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 395, in resolve [pipenv.exceptions.ResolutionFailure]: raise ResolutionFailure(message=str(e)) [pipenv.exceptions.ResolutionFailure]: pipenv.exceptions.ResolutionFailure: ERROR: ERROR: Could not find a version that matches mplfinance>=0.12 [pipenv.exceptions.ResolutionFailure]: Skipped pre-versions: 0.11.1a0, 0.11.1a0, 0.11.1a1, 0.11.1a1, 0.12.0a0, 0.12.0a0, 0.12.0a1, 0.12.0a1, 0.12.0a2, 0.12.0a2, 0.12.0a3, 0.12.0a3, 0.12.3a0, 0.12.3a0, 0.12.3a1, 0.12.3a1, 0.12.3a2, 0.12.3a2, 0.12.3a3, 0.12.3a3, 0.12.4a0, 0.12.4a0, 0.12.5a1, 0.12.5a1, 0.12.5a2, 0.12.5a2, 0.12.5a3, 0.12.5a3, 0.12.6a0, 0.12.6a0, 0.12.6a1, 0.12.6a1, 0.12.6a2, 0.12.6a2, 0.12.6a3, 0.12.6a3, 0.12.7a0, 0.12.7a0, 0.12.7a3, 0.12.7a3, 0.12.7a4, 0.12.7a4, 0.12.7a5, 0.12.7a5, 0.12.7a7, 0.12.7a7, 0.12.7a10, 0.12.7a10, 0.12.7a12, 0.12.7a12, 0.12.7a17, 0.12.7a17 [pipenv.exceptions.ResolutionFailure]: Warning: Your dependencies could not be resolved. You likely have a mismatch in your sub-dependencies. First try clearing your dependency cache with $ pipenv lock --clear, then try the original command again. Alternatively, you can use $ pipenv install --skip-lock to bypass this mechanism, then run $ pipenv graph to inspect the situation. Hint: try $ pipenv lock --pre if it is a pre-release dependency. ERROR: ERROR: Could not find a version that matches mplfinance>=0.12 Skipped pre-versions: 0.11.1a0, 0.11.1a0, 0.11.1a1, 0.11.1a1, 0.12.0a0, 0.12.0a0, 0.12.0a1, 0.12.0a1, 0.12.0a2, 0.12.0a2, 0.12.0a3, 0.12.0a3, 0.12.3a0, 0.12.3a0, 0.12.3a1, 0.12.3a1, 0.12.3a2, 0.12.3a2, 0.12.3a3, 0.12.3a3, 0.12.4a0, 0.12.4a0, 0.12.5a1, 0.12.5a1, 0.12.5a2, 0.12.5a2, 0.12.5a3, 0.12.5a3, 0.12.6a0, 0.12.6a0, 0.12.6a1, 0.12.6a1, 0.12.6a2, 0.12.6a2, 0.12.6a3, 0.12.6a3, 0.12.7a0, 0.12.7a0, 0.12.7a3, 0.12.7a3, 0.12.7a4, 0.12.7a4, 0.12.7a5, 0.12.7a5, 0.12.7a7, 0.12.7a7, 0.12.7a10, 0.12.7a10, 0.12.7a12, 0.12.7a12, 0.12.7a17, 0.12.7a17 There are incompatible versions in the resolved dependencies. ['Traceback (most recent call last):\n', ' File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 501, in create_spinner\n yield sp\n', ' File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 649, in venv_resolve_deps\n c = resolve(cmd, sp)\n', ' File "/usr/local/.pyenv/versions/3.9.4/lib/python3.9/site-packages/pipenv/utils.py", line 539, in resolve\n sys.exit(c.return_code)\n', 'SystemExit: 1\n'] ``` If you think the above is an error on Dependabot's side please don't hesitate to get in touch - we'll do whatever we can to fix it. [View the update logs](https://app.dependabot.com/accounts/twelvedata/update-logs/77915359).
1medium
Title: Create & Add another IP address on interfaces is not working correctly Body: ### Deployment Type Self-hosted ### Triage priority N/A ### NetBox Version V4.1.7 ### Python Version 3.12 ### Steps to Reproduce 1.add IP address to a interface.. 2. click create & add another. 3. add ip address details and click create. ### Expected Behavior should add second ip address to interface ### Observed Behavior pop says IP Address has been created but when go back to device interface tab the IP address is no where to be seen
1medium
Title: DoesNotExist at /en/software/features - Language matching query does not exist. Body: ## Steps to Reproduce I did a clean install, following the docker guide for setting up a docker production instance (used the default docker-compose). **Expected results:** enter ip:port and access the WebUI with default credentials **Actual results:** Continuously getting a django error <details> <summary>Logs</summary> <!-- Any logs you think would be useful (if you have a local instance) --> ```bash Environment: Request Method: GET Request URL: http://10.100.10.2/en/software/features Django Version: 3.2.10 Python Version: 3.9.5 Installed Applications: ('django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.messages', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.staticfiles', 'django_extensions', 'storages', 'wger.config', 'wger.core', 'wger.mailer', 'wger.exercises', 'wger.gym', 'wger.manager', 'wger.nutrition', 'wger.software', 'wger.utils', 'wger.weight', 'wger.gallery', 'wger.measurements', 'captcha', 'django.contrib.sitemaps', 'easy_thumbnails', 'compressor', 'crispy_forms', 'rest_framework', 'rest_framework.authtoken', 'django_filters', 'django_bootstrap_breadcrumbs', 'corsheaders') Installed Middleware: ('corsheaders.middleware.CorsMiddleware', 'django.middleware.common.CommonMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'wger.utils.middleware.JavascriptAJAXRedirectionMiddleware', 'wger.utils.middleware.WgerAuthenticationMiddleware', 'wger.utils.middleware.RobotsExclusionMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'django.middleware.locale.LocaleMiddleware') Traceback (most recent call last): File "/home/wger/src/wger/utils/language.py", line 54, in load_language language = Language.objects.get(short_name=used_language) File "/usr/local/lib/python3.9/dist-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/local/lib/python3.9/dist-packages/django/db/models/query.py", line 435, in get raise self.model.DoesNotExist( During handling of the above exception (Language matching query does not exist.), another exception occurred: File "/usr/local/lib/python3.9/dist-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/dist-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/home/wger/src/wger/software/views.py", line 37, in features return render(request, 'features.html', context) File "/usr/local/lib/python3.9/dist-packages/django/shortcuts.py", line 19, in render content = loader.render_to_string(template_name, context, request, using=using) File "/usr/local/lib/python3.9/dist-packages/django/template/loader.py", line 62, in render_to_string return template.render(context, request) File "/usr/local/lib/python3.9/dist-packages/django/template/backends/django.py", line 61, in render return self.template.render(context) File "/usr/local/lib/python3.9/dist-packages/django/template/base.py", line 168, in render with context.bind_template(self): File "/usr/lib/python3.9/contextlib.py", line 117, in __enter__ return next(self.gen) File "/usr/local/lib/python3.9/dist-packages/django/template/context.py", line 244, in bind_template updates.update(processor(self.request)) File "/home/wger/src/wger/utils/context_processor.py", line 28, in processor language = load_language() File "/home/wger/src/wger/utils/language.py", line 57, in load_language language = Language.objects.get(short_name="en") File "/usr/local/lib/python3.9/dist-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/local/lib/python3.9/dist-packages/django/db/models/query.py", line 435, in get raise self.model.DoesNotExist( Exception Type: DoesNotExist at /en/software/features Exception Value: Language matching query does not exist. ``` Attached a full report: [DoesNotExist at _en_software_features.pdf](https://github.com/wger-project/wger/files/7797751/DoesNotExist.at._en_software_features.pdf) </details> **At the bottom it is mentioned, that I am seeing this stack Trace because Django Debug was set to True. But the prod.env is set to False...**
2hard
Title: pro_bar获取不到50ETF数据 Body: Hi Jimmy, 在尝试用pro_bar获取50ETF数据(华夏基金的:510050.SH)时,返回为None 代码如下: pro = ts.pro_api('my token') df = ts.pro_bar(pro_api=pro, ts_code='510050.SH', adj='qfq', start_date='20180101', end_date='20181011',asset='fd') print(df) 返回为None (注意,asset='i','e','fd'都分别试过,都返回是None) 问题: 1、pro_bar是否可以获取510050.SH的数据? 2、如果可以,是否可以获取5分钟数据:freq=5min? 3、获取的数据(日线或者分钟数据),adj='qfq'是否有效?(现在我用的ts.bar来获取的,但是获取时adj='qfq'是没有用的。所以我不清楚pro_bar是否会修正这一点?) 谢谢
1medium
Title: Crashes when executing model quantification on Deeplabv3 Body: # Bug Report ### Is the issue related to model conversion? No ### Describe the bug I want to perform INT8 precision inference on the Deeplabv3 model on the CPU. I first quantified the model, but during execution, I threw NotImplemented: [ONNXRuntimeError] : 9 : NOT_IMPLEMENTED : Could not find an implementation for ConvInteger(10) node with name '/resnet/conv1/Conv_quant' Error. The model file is shown in the attachment (https://drive.google.com/file/d/10QhV6_lqoD4nnGx3a4HJV-zA7bQJ86E5/view?usp=drive_link). ### System information - OS Platform and Distribution (Linux Ubuntu 20.04): - ONNX version (1.17.0): - Python version:3.9.19 ### Reproduction instructions import onnxruntime as ort from onnxruntime.quantization import quantize_dynamic, QuantType import numpy as np session = ort.InferenceSession("./deeplabv3.onnx", providers=["CPUExecutionProvider"],graph_optimization_level=ort.GraphOptimizationLevel.ORT_ENABLE_EXTENDED) quantized_model = quantize_dynamic("./deeplabv3.onnx", "./deeplabv3_quantized.onnx", weight_type=QuantType.QInt8) input_data = np.random.rand(2, 3, 513, 513).astype(np.float32) input_data_int8 = input_data.astype(np.int8) inputs = {session.get_inputs()[0].name: input_data} outputs = session.run(["infer_output"], inputs) session_int8 = ort.InferenceSession("../deeplabv3_quantized.onnx", providers=["CPUExecutionProvider"]) outputs_int8 = session_int8.run(["infer_output"], quantized_model) ### Expected behavior Execute normally.
2hard
Title: History of executed notebooks Body: Get history of last executed notebooks
1medium
Title: Route order problem. Body: At present, we define route as follows: ```python route = {'/movies/?page=1': '/html/gndy/dyzz/', '/movies/?page=:page': '/html/gndy/dyzz/index_:page.html', '/movies/': '/html/gndy/dyzz/'} ``` The problem is the ordering. Use tuple or OrderedDict
0easy
Title: faa 命令行初始化项目时有点版本依赖的问题 Body: fastapi-user-auth 0.5.0 requires fastapi-amis-admin<0.6.0,>=0.5.0, but you have fastapi-amis-admin 0.6.1 which is incompatible.
2hard
Title: How to use jwt tokens in django-oauth-toolkit without adding to the database? Body: I have been working on separating the authorization server and the resource server using django-oauth-toolkit. Also I wrote my jwt token generator: ``` from rest_framework_simplejwt.tokens import RefreshToken def my_acces_token_generator (request, refresh_token = False): refresh = RefreshToken.for_user (request.user) return str (refresh.access_token) def my_refresh_token_generator (request, refresh_token = False): refresh = RefreshToken.for_user (request.user) return str (refresh) ``` And now I noticed that my jwt tokens are added to the database created by django-oauth-toolkit, but the essence of jwt is not to store them in the database. Tell me how to do it all rummaged through I can not find anything. ![Снимок](https://user-images.githubusercontent.com/45039914/141380761-01edc794-ffb2-4e4d-8694-f6d04dac08e0.PNG)
1medium
Title: AttributeError: 'WorkerController' object has no attribute 'slaveinput' Body: I recently upgraded from pytest-xdist 1.34.0 to 2.0.0 and found that tests always fail with the new version. I pinned the previous version as a workaround. I've tested on Ubuntu 18.04 and the circle [CI docker image for python 3.7.6](https://github.com/CircleCI-Public/circleci-dockerfiles/blob/master/python/images/3.7.6/Dockerfile) If I remove --cov=project_folder (project_folder was redacted), the tests run successfully. ``` pytest --numprocesses=auto --cov=project_folder test itest ============================= test session starts ============================== platform linux -- Python 3.7.6, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /root/repo plugins: xdist-2.0.0, forked-1.3.0, cov-2.10.0 gw0 C / gw1 I / gw2 I / gw3 I / gw4 I / gw5 I / gw6 I / gw7 I / gw8 I / gw9 I / gw10 I / gw11 I / gw12 I / gw13 I / gw14 I / gw15 I / gw16 I / gw17 IINTERNALERROR> Traceback (most recent call last): INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/_pytest/main.py", line 238, in wrap_session INTERNALERROR> config.hook.pytest_sessionstart(session=session) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/hooks.py", line 286, in __call__ INTERNALERROR> return self._hookexec(self, self.get_hookimpls(), kwargs) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/manager.py", line 93, in _hookexec INTERNALERROR> return self._inner_hookexec(hook, methods, kwargs) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/manager.py", line 87, in <lambda> INTERNALERROR> firstresult=hook.spec.opts.get("firstresult") if hook.spec else False, INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 208, in _multicall INTERNALERROR> return outcome.get_result() INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 80, in get_result INTERNALERROR> raise ex[1].with_traceback(ex[2]) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 187, in _multicall INTERNALERROR> res = hook_impl.function(*args) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/xdist/dsession.py", line 78, in pytest_sessionstart INTERNALERROR> nodes = self.nodemanager.setup_nodes(putevent=self.queue.put) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/xdist/workermanage.py", line 65, in setup_nodes INTERNALERROR> return [self.setup_node(spec, putevent) for spec in self.specs] INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/xdist/workermanage.py", line 65, in <listcomp> INTERNALERROR> return [self.setup_node(spec, putevent) for spec in self.specs] INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/xdist/workermanage.py", line 73, in setup_node INTERNALERROR> node.setup() INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/xdist/workermanage.py", line 260, in setup INTERNALERROR> self.config.hook.pytest_configure_node(node=self) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/hooks.py", line 286, in __call__ INTERNALERROR> return self._hookexec(self, self.get_hookimpls(), kwargs) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/manager.py", line 93, in _hookexec INTERNALERROR> return self._inner_hookexec(hook, methods, kwargs) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/manager.py", line 87, in <lambda> INTERNALERROR> firstresult=hook.spec.opts.get("firstresult") if hook.spec else False, INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 208, in _multicall INTERNALERROR> return outcome.get_result() INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 80, in get_result INTERNALERROR> raise ex[1].with_traceback(ex[2]) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pluggy/callers.py", line 187, in _multicall INTERNALERROR> res = hook_impl.function(*args) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pytest_cov/plugin.py", line 239, in pytest_configure_node INTERNALERROR> self.cov_controller.configure_node(node) INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pytest_cov/engine.py", line 274, in configure_node INTERNALERROR> workerinput(node).update({ INTERNALERROR> File "/root/repo/venv/lib/python3.7/site-packages/pytest_cov/compat.py", line 42, in fn INTERNALERROR> return getattr(obj, attr, *args) INTERNALERROR> AttributeError: 'WorkerController' object has no attribute 'slaveinput' ```
2hard
Title: New st.query_params doesn't persist state in URL after page refresh while deprecated st.experimental_set_query_params does Body: ### Checklist - [x] I have searched the [existing issues](https://github.com/streamlit/streamlit/issues) for similar issues. - [x] I added a very descriptive title to this issue. - [x] I have provided sufficient information below to help reproduce this issue. ### Summary The new `st.query_params` API does not maintain state in URL parameters after page refresh, while the deprecated `st.experimental_get/set_query_params` methods work correctly. This is blocking migration from experimental to stable API for applications that rely on URL parameters for state management. ### Reproducible Code Example ```Python ########################### # login_experimental_api.py ########################### import streamlit as st st.set_page_config(page_title="Login State Test - Experimental API") # Initialize login state if "is_logged" not in st.session_state: query_params = st.experimental_get_query_params() st.session_state.is_logged = query_params.get("logged_in", ["False"])[0] == "True" # Display current state st.write("Current login state:", st.session_state.is_logged) st.write("Current query parameters:", st.experimental_get_query_params()) if st.button("Login"): st.session_state.is_logged = True st.experimental_set_query_params(logged_in="True") st.rerun() st.markdown("---") st.markdown(""" ### How to test: 1. Click "Login" - you'll see the state change to logged in 2. Note the URL parameters change 3. Refresh the page - the state persists 4. Check URL parameters - they are maintained This demonstrates that the deprecated `st.experimental_set_query_params` maintains state in URL parameters after page refresh. """) ################## # login_new_api.py ################## import streamlit as st st.set_page_config(page_title="Login State Test - New API") # Initialize login state if "is_logged" not in st.session_state: st.session_state.is_logged = st.query_params.get("logged_in", ["False"])[0] == "True" # Display current state st.write("Current login state:", st.session_state.is_logged) st.write("Current query parameters:", dict(st.query_params)) if st.button("Login"): st.session_state.is_logged = True st.query_params.logged_in = "True" st.rerun() st.markdown("---") st.markdown(""" ### How to test: 1. Click "Login" - you'll see the state change to logged in 2. Note the URL parameters change 3. Refresh the page - the state is lost 4. Check URL parameters - they are also lost This demonstrates that the new recommended `st.query_params` API doesn't maintain state in URL parameters after page refresh. """) ``` ### Steps To Reproduce 1. Run the attached minimal example files 2. Click "Login" - state will persist after refresh in `login_experimental_api.py` 3. Click "Login" - state will be lost after refresh in `login_new_api.py` ### Expected Behavior `streamlit run login_new_api.py` should keep the session. ### Current Behavior Using old API Streamlit page shows yellow alerts: ``` Please replace st.experimental_get_query_params with st.query_params. st.experimental_get_query_params will be removed after 2024-04-11. Refer to our [docs page](https://docs.streamlit.io/develop/api-reference/caching-and-state/st.query_params) for more information. ``` Whilst terminal throws deprecation warnings with the same content: ``` $ streamlit run login_experimental_api.py You can now view your Streamlit app in your browser. Local URL: http://localhost:8501 Network URL: http://10.0.100.20:8501 For better performance, install the Watchdog module: $ xcode-select --install $ pip install watchdog 2025-02-15 01:25:14.734 Please replace `st.experimental_get_query_params` with `st.query_params`. `st.experimental_get_query_params` will be removed after 2024-04-11. Refer to our [docs page](https://docs.streamlit.io/develop/api-reference/caching-and-state/st.query_params) for more information. 2025-02-15 01:25:14.734 Please replace `st.experimental_get_query_params` with `st.query_params`. `st.experimental_get_query_params` will be removed after 2024-04-11. Refer to our [docs page](https://docs.streamlit.io/develop/api-reference/caching-and-state/st.query_params) for more information. ``` ### Is this a regression? - [x] Yes, this used to work in a previous version. ### Debug info - Streamlit version: `Streamlit, version 1.39.0` - Python version: `Python 3.12.9` - Operating System: `macOS 15.2` - Browser: Brave Browser, Safari, Google Chrome ``` ❯ streamlit version Streamlit, version 1.39.0 ❯ python --version Python 3.12.9 ❯ sw_vers ProductName: macOS ProductVersion: 15.2 BuildVersion: 24C101 ``` ### Additional Information _No response_
2hard
Title: Unable to build from PyPi source file due to missing `build_hook.py` Body: ### Describe the bug Building with [source](https://pypi.org/project/marimo/#marimo-0.11.0.tar.gz) distributed by PyPi will result in an error in the `0.11.0` release: <details> ``` ❯ python -m build --wheel --no-isolation * Getting build dependencies for wheel... Traceback (most recent call last): File "/usr/lib/python3.13/site-packages/pyproject_hooks/_in_process/_in_process.py", line 389, in <module> main() ~~~~^^ File "/usr/lib/python3.13/site-packages/pyproject_hooks/_in_process/_in_process.py", line 373, in main json_out["return_val"] = hook(**hook_input["kwargs"]) ~~~~^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.13/site-packages/pyproject_hooks/_in_process/_in_process.py", line 143, in get_requires_for_build_wheel return hook(config_settings) File "/usr/lib/python3.13/site-packages/hatchling/build.py", line 44, in get_requires_for_build_wheel return builder.config.dependencies ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.13/functools.py", line 1039, in __get__ val = self.func(instance) File "/usr/lib/python3.13/site-packages/hatchling/builders/config.py", line 577, in dependencies for dependency in self.dynamic_dependencies: ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/lib/python3.13/functools.py", line 1039, in __get__ val = self.func(instance) File "/usr/lib/python3.13/site-packages/hatchling/builders/config.py", line 593, in dynamic_dependencies build_hook = build_hook_cls( self.root, config, self, self.builder.metadata, '', self.builder.PLUGIN_NAME, self.builder.app ) File "/usr/lib/python3.13/site-packages/hatchling/builders/hooks/custom.py", line 33, in __new__ raise OSError(message) OSError: Build script does not exist: build_hook.py ERROR Backend subprocess exited when trying to invoke get_requires_for_build_wheel ``` </details> This is caused by the change in [commit 52906ba](https://github.com/marimo-team/marimo/commit/52906baa3fae4907711af82ad69a8866408e94b5#diff-50c86b7ed8ac2cf95bd48334961bf0530cdc77b5a56f852c5c61b89d735fd711) to `pyproject.toml`: ``` ... [tool.hatch.build.hooks.custom] path = "build_hook.py" ... ``` It can be fixed by placing the `build_hook.py` from the Github repo into the extracted source folder. ### Environment <details> ``` { "marimo": "0.11.0", "OS": "Linux", "OS Version": "6.12.10-zen1-1-zen", "Processor": "", "Python Version": "3.13.1", "Binaries": { "Browser": "--", "Node": "v23.7.0" }, "Dependencies": { "click": "8.1.7", "docutils": "0.21.2", "itsdangerous": "2.1.2", "jedi": "0.19.2", "markdown": "3.7", "narwhals": "1.24.2", "packaging": "24.2", "psutil": "6.1.1", "pygments": "2.19.1", "pymdown-extensions": "10.14.1", "pyyaml": "6.0.2", "ruff": "missing", "starlette": "0.45.3", "tomlkit": "0.13.2", "typing-extensions": "4.12.2", "uvicorn": "0.34.0", "websockets": "12.0" }, "Optional Dependencies": {}, "Experimental Flags": {} } ``` </details> ### Code to reproduce _No response_
1medium
Title: Create infinite stream from image Body: Hello! can someone tell me how i can make an endless stream from one picture? this picture may change over time there is my code: ``` def loop_image(): img = cv2.imread('./images/1.jpg') img = cv2.resize(img, (640, 480)) new_frame = VideoFrame.from_ndarray(img, format="bgr24") return new_frame class VideoStreamTrack(VideoStreamTrack): def __init__(self): super().__init__() # don't forget this! self.counter = 0 self.frames = [] for i in range(3): print(f"append image {i}") self.frames.append(loop_image()) async def recv(self): print("append image") pts, time_base = await self.next_timestamp() frame = self.frames[self.counter % 30] frame.pts = pts frame.time_base = time_base self.counter += 1 self.frames.append(loop_image()) return frame ```
1medium
Title: Arguments and vertical Japanese text Body: I have used EasyOCR successfully a few times now, however I'm running into issues. Firstly, a guide recommended using a "paragraph" argument. Putting "--paragraph=1" into my script caused an error called "Unrecognized arguments" and that error persisted on every variation I tried. Typing "easyocr -h" tells me that the options I can use are language, input file, detail of output and if I want to use GPU. These options seem very bare bones and in contrast to guides. Also, Japanese vertical text does not work. It only grabs the uppermost character of each column and puts it on its own line in the output (which is very annoying even when it does manage to grab the text correctly). Can that be fixed? I tried downloading the bigger japanese.pth and replacing the one in the install directory, but the program deleted it. Is it not compatible with the latest git download? Did I get the wrong version? Is there even a way to check version? The standard "--version" did absolutely nothing. One last thing, how do I make it stop pestering me about my two different GPUs? To clarify, I'm using Windows 10's CMD to run EasyOCR.
1medium
Title: Comprehensive Documentation of Supported Input Formats Body: ## Goal To provide clear and comprehensive documentation detailing the support status for various input formats like JsonSchema and OpenAPI, including different drafts like Draft 7 and Draft 2019-09, to help users understand the extent to which they can utilize datamodel-code-generator with different specifications. ## Background The current documentation needs more detailed information on the support status of various input formats. Users may need help understanding what aspects of JsonSchema or OpenAPI are supported, especially concerning different drafts. This lack of clarity could hinder the utilization of datamodel-code-generator to its full extent. ## Suggested Documentation Structure: ### Supported Input Formats: List of all supported input formats with a brief description. #### JsonSchema: Support status for Draft 7, Draft 2019-09, etc. Any known limitations or workarounds. ##### OpenAPI: Support status for versions 3.0, 3.1, etc. Any known limitations or workarounds. ## Related Issues https://github.com/koxudaxi/datamodel-code-generator/issues/1578
1medium
Title: receiving error with face_encoding of face_recognition library after cuda setup Body: * face_recognition version: 1.2.3 * Python version: 3.6.5 * Operating System: windows 10 ### What I Did I have just setup my CUDA with dlib and after that I am receiving this error when running the encoding of face using face_recognition library ``` face_encodings = face_recognition.face_encodings(frame, face_locations, num_jitters=1) File "C:\Users\abhinav.jhanwar\AppData\Local\Continuum\anaconda3\lib\site-packages\face_recognition\api.py", line 210, in face_encodings :return: A list of 128-dimensional face encodings (one for each face in the image) File "C:\Users\abhinav.jhanwar\AppData\Local\Continuum\anaconda3\lib\site-packages\face_recognition\api.py", line 210, in <listcomp> :return: A list of 128-dimensional face encodings (one for each face in the image) RuntimeError: Error while calling cudnnCreate(&handles[new_device_id]) in file C:\dlib-master\dlib-master\dlib\cuda\cudnn_dlibapi.cpp:104. code: 1, reason: CUDA Runtime API initialization failed. ```
2hard
Title: Document that failing test setup stops execution even if continue-on-failure mode is active Body: When support for controlling the continue-on-failure mode was added (#2285), we didn't think how tests setups should behave when the mode is activated. Currently if you have a test like ```robotframework *** Test Cases *** Example [Tags] robot:continue-on-failure [Setup] Setup Keyword Keyword In Body Another Keyword ``` and the `Setup Keyword` fails, keywords in the test body aren't run. That can be considered inconsistent because continue-on-failure mode is explicitly enabled, but I actually like this behavior. You can have some action in setup so that execution ends if it fails, but keywords in the body that are validating that everything went as expected would all be run if the setup succeeds. This would also add some new functionality to setups that are otherwise not really different from normal keywords. What do others (especially @oboehmer) think about this? Should we stop execution after a failing setup also when continue-on-failure mode is on? If yes, we just need to add some tests for the current behavior and mention this in the User Guide. If not, then we need to change the code as well. A problem with changing the code is that the change would be backwards incompatible. It's a bit questionable could we do it without a deprecation in RF 5.1 but deprecation adds more work.
1medium
Title: d2l.Seq2SeqEncoder(vocab_size=10, embed_size=8, num_hiddens=16, 2 num_layers=2) Body: --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[21], [line 4](vscode-notebook-cell:?execution_count=21&line=4) [1](vscode-notebook-cell:?execution_count=21&line=1) encoder = d2l.Seq2SeqEncoder(vocab_size=10, embed_size=8, num_hiddens=16, [2](vscode-notebook-cell:?execution_count=21&line=2) num_layers=2) [3](vscode-notebook-cell:?execution_count=21&line=3) encoder.eval() ----> [4](vscode-notebook-cell:?execution_count=21&line=4) decoder = Seq2SeqAttentionDecoder(vocab_size=10, embed_size=8, num_hiddens=16, [5](vscode-notebook-cell:?execution_count=21&line=5) num_layers=2) [6](vscode-notebook-cell:?execution_count=21&line=6) decoder.eval() [7](vscode-notebook-cell:?execution_count=21&line=7) X = torch.zeros((4, 7), dtype=torch.long) # (batch_size,num_steps) Cell In[20], [line 5](vscode-notebook-cell:?execution_count=20&line=5) [2](vscode-notebook-cell:?execution_count=20&line=2) def __init__(self, vocab_size, embed_size, num_hiddens, num_layers, [3](vscode-notebook-cell:?execution_count=20&line=3) dropout=0, **kwargs): [4](vscode-notebook-cell:?execution_count=20&line=4) super(Seq2SeqAttentionDecoder, self).__init__(**kwargs) ----> [5](vscode-notebook-cell:?execution_count=20&line=5) self.attention = d2l.AdditiveAttention( [6](vscode-notebook-cell:?execution_count=20&line=6) num_hiddens, num_hiddens, num_hiddens, dropout) [7](vscode-notebook-cell:?execution_count=20&line=7) self.embedding = nn.Embedding(vocab_size, embed_size) [8](vscode-notebook-cell:?execution_count=20&line=8) self.rnn = nn.GRU( [9](vscode-notebook-cell:?execution_count=20&line=9) embed_size + num_hiddens, num_hiddens, num_layers, [10](vscode-notebook-cell:?execution_count=20&line=10) dropout=dropout) TypeError: __init__() takes 3 positional arguments but 5 were given
1medium
Title: Feature Request: SRT (Subtitle) Support Body: (I can't do this as a pull request, I'm not a Python programmer.) I would like to request that the translate file mode support SRT (Subtitle) files. The format of these is pretty simple, they are text files with an incrementing number, the next line has a time range in the form of `HH:MM:SS,HUN --> HH:MM:SS,HUN`, and finally the text which **may** include some HTML markup, primarily either `<b>` or` <i>`. Sometimes there is a `<br>` (or `<br/>`). A blank line separates text 'frames'. Mostly LibreTranslate needs to _ignore_ the number and the time ranges entirely, and for the text, treat it like HTML. There may already be Python code somewhere for handling SRT files. ### Example: ``` 1 00:00:02,830 --> 00:00:05,560 <i>The miniature giant space hamster descends from the starry sky.</i> 2 00:00:05,700 --> 00:00:06,360 Boo! ``` (If someone decides to work on this and needs sample SRTs, please contact me.) ### Use Case: Translating subtitle text is an obvious use of LibreTranslate. For example, _Subtitle Edit_, an application to aid in editing or creating subtitle files, even has a "Google Translate" button with in it for machine translating lines of dialogue (it just opens a web browser). ### Notes: At present LibreTranslate handles srt (as txt) in kind of a mixed bag. Sometimes it trims or strips out the time range and sometimes it changes the frame number although it's incredibly inconsistent on both.
1medium
Title: BP008: Not Handling conint(ge= , lt=) Correctly Body: Python version: 3.10.12 bump-pydantic version: 0.6.1 Pydantic version: 2.3.0 I installed and ran `bump-pydantic` per the project's README.md file instructions against my FastAPI project[*](https://github.com/questionlp/api.wwdt.me_v2). The resulting diff doesn't handle `conint(ge=, lt=)` correctly. Before running `bump-pydantic`, one of the models looks like: ```python class GuestAppearance(BaseModel): """Appearance Information""" show_id: conint(ge=0, lt=2**31) = Field(title="Show ID") ``` The diff generated by `bump-pydantic` returns: ```python class GuestAppearance(BaseModel): """Appearance Information""" show_id: Annotated = Field(title="Show ID") ``` It is losing both the `int` type and the `ge` and `lt` constraints. Replacing the expression `2**31` with `2147483648` doesn't change how `bump-pydantic` handles the `conint` per the following diff: ``` class GuestAppearance(BaseModel): """Appearance Information""" - show_id: conint(ge=0, lt=2147483648) = Field(title="Show ID") + show_id: Annotated = Field(title="Show ID") ```
1medium
Title: Having a hard time getting set up on Ubuntu 24.04 Body: The output of `thefuck --version` (something like `The Fuck 3.1 using Python 3.5.0 and Bash 4.4.12(1)-release`): ``` pblanton@ThreadRipper:~$ thefuck --version Traceback (most recent call last): File "/home/pblanton/.local/bin/thefuck", line 5, in <module> from thefuck.entrypoints.main import main File "/home/pblanton/.local/lib/python3.12/site-packages/thefuck/entrypoints/main.py", line 8, in <module> from .. import logs # noqa: E402 ^^^^^^^^^^^^^^^^^^^ File "/home/pblanton/.local/lib/python3.12/site-packages/thefuck/logs.py", line 8, in <module> from .conf import settings File "/home/pblanton/.local/lib/python3.12/site-packages/thefuck/conf.py", line 1, in <module> from imp import load_source ModuleNotFoundError: No module named 'imp' ``` Your system (Debian 7, ArchLinux, Windows, etc.): Ubuntu 24.04 How to reproduce the bug: install on Ubuntu 24.04 and try to run `thefuck --version` The output of The Fuck with `THEFUCK_DEBUG=true` exported (typically execute `export THEFUCK_DEBUG=true` in your shell before The Fuck): Same error
1medium
Title: Brush labels to YOLO format Body: ### Search before asking - [x] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/orgs/ultralytics/discussions) and found no similar questions. ### Question Hello! Recenty I have used SAM model in label studio to assist my labeling task. It works pretty nice. However, I met some problems in exporting the dataset. It can't be exported in COCO or YOLO format since SAM create brush labels. Could you tell me how to convert my brush labels to a yolo format dataset? ### Additional _No response_
1medium
Title: Add unit test for crypto.SecureTLSConfig to double check on go version bumps Body: SecureTLSConfig specifies preferred ciphers for serving. When we switch go versions, a unit test should inform us when more secure suites are available, so we remember to add them if desired
1medium
Title: New Example, Porting a simple two-tier app to containers and Kubernetes Body: This example starts with a simple two-tier guestbook app using MySQL. It then shows the steps taken to first port the app to run in Docker containers, and then port to Kubernetes. Lastly it covers setting the MySQL password using the Kubernetes secrets feature.
1medium
Title: ObjectType resolving fields from dicts or objects Body: Hi! There is an example code: ``` class Photo(graphene.ObjectTypet): id = graphene.ID() url = graphene.String() fileName = graphene.String() extension = graphene.String() class User(graphene.ObjectTypet): _id = graphene.ID() name = graphene.String() photos = graphene.List(Photo) class Query(graphene.ObjectType): user = graphene.Field(User, dict(_id=graphene.String(required=True))) async def resolve_user(self, info, _id): user = await db.users.find_one(dict(_id=_id)) # do the simple filted resolved_fileds = {f: user[f] for f in User._meta.fields} user = User(**resolved_fileds) if user else None return user ``` The result is ``` { "data": { "user": { "Id": "1", "name": "Test User", "photos": [ { "id": null, "fileName": null } ] } } } ``` Field "photos" not initialized in __init__. In argument "photos" I got just list with dicts. I know that I can to this serialization by myself, but maybe it's already implemented?
1medium
Title: How to use python-gitlab library to search a string in every commits? Body: I noticed `commits = project.commits.list(all=True)` can list every commits, but I don't know how to perform a search against each commits, can it be done? :)
1medium
Title: Downloading Linux Agent without Code Signing Body: **Server Info (please complete the following information):** - OS: Ubuntu Server 20.04 - Browser: vivaldi (chromium) - RMM Version 0.13.3 **Installation Method:** - [x] Standard - [ ] Docker **Agent Info (please complete the following information):** - Agent version (as shown in the 'Summary' tab of the agent from web UI): - Agent OS: 2.0.3 **Describe the bug** Whenever I try to download any linux agent over the top bar menu, I get a small red error notification in the top middle (Missing code signing token - 400: Bad Request). Since permanent links are just usable for windows, I am not able to get any deployment link at all. I have tried all combinations of server and architecture. **To Reproduce** Steps to reproduce the behavior: 1. Go to dashboard 2. Click on Agents (top bar) 3. Click on Install Agent 4. Select any Linux based configuration, any Client/Site Combination, expiration, server/workstation.... 5. Click on Download Linux Install Script 6. See the error **Expected behavior** I expected the linux install script to download just as the windows installer which works perfectly fine. **Screenshots** ![image](https://user-images.githubusercontent.com/80537345/166639892-909d8d9e-6a48-4671-9801-c717225ae554.png) **Additional context** Add any other context about the problem here.
1medium