text
stringlengths 20
57.3k
| labels
class label 4
classes |
---|---|
Title: Add timezone property to slack_sdk.models.TimePickerElement class
Body: While Node and Java SDKs already have the new property, this Python SDK still does not have the newly added `timezone` property in the [`TimePickerElement`](https://github.com/slackapi/python-slack-sdk/blob/v3.18.1/slack_sdk/models/blocks/block_elements.py#L454-L511).
We are still waiting for the document updates [here](https://api.slack.com/reference/block-kit/block-elements#timepicker) but we can add the property with some general description (we can update it once the API documents are ready).
See also:
* https://github.com/slackapi/node-slack-sdk/issues/1502
* https://github.com/slackapi/node-slack-sdk/pull/1503
* https://github.com/slackapi/java-slack-sdk/pull/1010
### Category (place an `x` in each of the `[ ]`)
- [ ] **slack_sdk.web.WebClient (sync/async)** (Web API client)
- [ ] **slack_sdk.webhook.WebhookClient (sync/async)** (Incoming Webhook, response_url sender)
- [x] **slack_sdk.models** (UI component builders)
- [ ] **slack_sdk.oauth** (OAuth Flow Utilities)
- [ ] **slack_sdk.socket_mode** (Socket Mode client)
- [ ] **slack_sdk.audit_logs** (Audit Logs API client)
- [ ] **slack_sdk.scim** (SCIM API client)
- [ ] **slack_sdk.rtm** (RTM client)
- [ ] **slack_sdk.signature** (Request Signature Verifier)
### Requirements
Please read the [Contributing guidelines](https://github.com/slackapi/python-slack-sdk/blob/main/.github/contributing.md) and [Code of Conduct](https://slackhq.github.io/code-of-conduct) before creating this issue or pull request. By submitting, you are agreeing to those rules.
| 0easy
|
Title: Docs upgrade
Body: Leaving this from #653, which is a pretty good thing!
If you're thinking about potential changes to the docs, one thing I've been thinking about is providing an alternate/additional view for users to discover functions. I think a user looking at our API pages can see all the functions `janitor` has, and what they do, but we could also have a version where someone is browsing the page, or coming from a search engine with the question,
*"How do I remove duplicate columns in pandas?"*
and can then find the function they want, potentially see what it would take in pure `pandas`, and how we do it in `pyjanitor`.
_Originally posted by @hectormz in https://github.com/ericmjl/pyjanitor/pull/653#issuecomment-628960074_ | 0easy
|
Title: Link to relationship
Body: ### Checklist
- [X] There are no similar issues or pull requests for this yet.
### Is your feature related to a problem? Please describe.
Add auto link in relationship. As example: https://python-sqladmin.herokuapp.com/admin/address/details/1 to get user I need to go to table users, find user(no search). Please make user_id a link to users
### Describe the solution you would like.
user_id is a link to users
### Describe alternatives you considered
_No response_
### Additional context
_No response_ | 0easy
|
Title: Replace all tb.ref examples in docs with tb.get
Body: With #92, a new way of creating reference objects was introduced.
- We will need to replace all instances of `tb.ref` with `tb.get` and also mention that one can use the indexing pattern to create the objects.
- Also mention that when the object is JSON serializable, when `tb.get` is called on that variable, the actual value will be fetched, and not the reference object, which was the case before. (This is the actual reason why we need to move away from `tb.ref` examples because they no longer return "references" always.)
- Provide a few code examples to demonstrate the new pattern.
[This](https://github.com/nteract/testbook/blob/main/testbook/client.py#L34-L51) piece of code outlines the different ways we can create testbook reference objects as of now. | 0easy
|
Title: [SDK] Show Valid Katib UI Link
Body: Check this thread: https://github.com/kubeflow/katib/pull/2098#discussion_r1085608043.
Currently, we show Katib UI link when Katib SDK is used from Kubernetes pod and SDK loads [`in-cluster` config](https://github.com/kubeflow/katib/blob/6bcbd2585198d85de90cbd7ffce89cc8bd7358cb/sdk/python/v1beta1/kubeflow/katib/api/katib_client.py#L57).
We use it to identify whether Katib SDK is ran from Kubeflow Notebook and we generate link to the appropriate Kubeflow Central Dashboard page.
User might also run Katib SDK `in-cluster`, but not from Kubeflow Notebooks UI.
We should find another approach to identify if Katib SDK is used from the Kubeflow Notebooks.
cc @kimwnasptd
/area sdk
---
<!-- Don't delete this message to encourage users to support your issue! -->
Love this feature? Give it a ๐ We prioritize the features with the most ๐
| 0easy
|
Title: Include `Message.html` in JSON results only if it is `True`
Body: Currently `robot.result.Message` is serialized to JSON so that the `html` attribute is always included. Including it only if it is `True` makes the resulting JSON data a bit smaller. That is also consistent with other similar attributes that are serialized to JSON only when their value differs from the default value.
Robot itself can load JSON where `html` is included, regardless is it `True` or `False`, and where it is omitted (in which case it is considered to be `False`) so in that regard the change is fully backwards compatible. External tools processing JSON results need to be updated if they expect `html` to be always set, though. JSON result support is so new that I don't expect many tools to be affected, so the change ought to be fine even in a non-major release. There are likely more tools like that in the future, especially if we get #3423 done, and delaying this change could thus cause bigger problems later. | 0easy
|
Title: Add welcome message for new members.
Body: It will be nice to make this bot behave more humanely by using this feature.
You'll add a new env variable for welcome message.
The bot DM every new member with the message.
It should be an embed in case someone wants style.
Happy new year. | 0easy
|
Title: FastAPI background_task cleanup is not executed upon error
Body: FastAPI `background_task` cleanup is not executed if any error is encountered.
To reproduce, run this code and add a print in cleanup function.
```py
import litserve as ls
class SimpleLitAPI(ls.LitAPI):
def setup(self, device):
# Set up the model, so it can be called in `predict`.
self.model = lambda x: x**2+"e"
def decode_request(self, request):
# Convert the request payload to your model input.
return request["input"]
def predict(self, x):
# Run the model on the input and return the output.
return self.model(x)
def encode_response(self, output):
# Convert the model output to a response payload.
return {"output": output}
``` | 0easy
|
Title: [FEATURE] Switch to `ruff`
Body: Might also be good to see what's up with the CI. It seems like the style checks are skipped? | 0easy
|
Title: Implement method __iter__ for Application object to avoid infinite loop "for a in app"
Body: The issue has been reported here: https://stackoverflow.com/a/41592702/3648361
The code to reproduce is very simple:
```python
from pywinauto import Application
app = Application()
for a in app:
print(a)
``` | 0easy
|
Title: Table: add css class to column headers
Body: ### Description
It would be helpful to have CSS classes associated to column header.
### Solution Proposed
I propose that the classes would be name via this procedure:
- prefix: taipy-table-header-
- suffix: the column name transformed to only allow A-Za-z\-_0-9
Make sure that the CSS class name is unique among all the columns of the table.
That calculation could be done completely in the front end.
TableUtils.ts seems a good place for this.
### Impact of Solution
Should be None
### Additional Context
_No response_
### Acceptance Criteria
- [X] Ensure new code is unit tested, and check code coverage is at least 90%.
- [X] Create related issue in taipy-doc for documentation and Release Notes.
- [X] Check if a new demo could be provided based on this, or if legacy demos could be benefit from it.
- [X] Ensure any change is well documented.
### Code of Conduct
- [x] I have checked the [existing issues](https://github.com/Avaiga/taipy/issues?q=is%3Aissue+).
- [ ] I am willing to work on this issue (optional) | 0easy
|
Title: Conversion Rate metric API
Body: The canonical definition is here: https://chaoss.community/?p=4924 | 0easy
|
Title: Docs: Wrong title for `concurrency` reference docs
Body: ### Summary
https://github.com/litestar-org/litestar/blame/7c4dd4bfceea7361ee76815550bfbf6cf185c5b8/docs/reference/concurrency.rst#L1 | 0easy
|
Title: Switch to databricks-sql-python library for SQL Integration
Body: ### ๐ The feature
We are currently using the sqlalchemy-databricks library for SQL integration in our project. While it has served us well, we believe that switching to the databricks-sql-python library would be beneficial for several reasons.
## Proposed Changes:
We propose replacing the existing sqlalchemy-databricks library with the databricks-sql-python library for SQL integration.
## Updated Features: The databricks-sql-python library may offer new features, improvements, and bug fixes that are not available in the older sqlalchemy-databricks library.
Community Adoption: Given the official support, the databricks-sql-python library is likely to have a larger and more active user community, which can be helpful for troubleshooting and finding resources.
# Steps to Implement:
- Replace the current sqlalchemy-databricks library with the databricks-sql-python library in our project's dependencies.
- Update any code that relies on the old library to use the new library's syntax and features.
- Test the project thoroughly to ensure that the integration with Databricks clusters works as expected.
### Motivation, pitch
Official Databricks Support: The databricks-sql-python library is officially supported by Databricks, which can provide better long-term stability and compatibility with Databricks clusters.
### Alternatives
If there are any additional details, considerations, or potential challenges related to this library switch, please provide them here.
### Additional context
_No response_ | 0easy
|
Title: Fix capitalisation of python (prefer Python) as per Vale suggestion
Body: Vale recommends we use "Python" rather than "python".
<img width="620" alt="image" src="https://github.com/user-attachments/assets/cc1e51ce-3dfc-44f8-b466-0e30d60bc06b">
Task: Search across Vizro docs (vizro-core and vizro-ai) for instances of python (not Python) and replace with discretion.
If there are genuine cases where lower case `python` is more appropriate, use the following to switch Vale off/on in situ
```markdown
<!-- vale off -->
Legitimate use of lower case p ("python")
<!--vale on-->
``` | 0easy
|
Title: The "item-pipeline" section example of the docs was not updated.
Body: <!--
Thanks for taking an interest in Scrapy!
If you have a question that starts with "How to...", please see the Scrapy Community page: https://scrapy.org/community/.
The GitHub issue tracker's purpose is to deal with bug reports and feature requests for the project itself.
Keep in mind that by filing an issue, you are expected to comply with Scrapy's Code of Conduct, including treating everyone with respect: https://github.com/scrapy/scrapy/blob/master/CODE_OF_CONDUCT.md
The following is a suggested template to structure your issue, you can find more guidelines at https://doc.scrapy.org/en/latest/contributing.html#reporting-bugs
-->
### Description
The [Take screenshot of item](https://docs.scrapy.org/en/2.10/topics/item-pipeline.html?highlight=spider.crawler.engine.download#take-screenshot-of-item) section in the document has not been updated, and it is still an example of version 2.9.0. You need to delete the "spider" parameter from the "item-pipeline.rst" document.
related issues: [issue 5994](https://github.com/scrapy/scrapy/issues/5994), [issue 5998](https://github.com/scrapy/scrapy/issues/5998).
### Steps to Reproduce
1. Run the "Take screenshot of item" example in the docs.
**Expected behavior:** The program is running fine.
**Actual behavior:** program error.
**Reproduces how often:** always.
### Versions
Please paste here the output of executing `scrapy version --verbose` in the command line.
```
Scrapy : 2.10.0
lxml : 4.9.2.0
libxml2 : 2.9.14
cssselect : 1.2.0
parsel : 1.8.1
w3lib : 2.1.1
Twisted : 22.10.0
Python : 3.8.5 (default, Jan 17 2023, 13:24:51) - [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)]
pyOpenSSL : 23.2.0 (OpenSSL 3.1.1 30 May 2023)
cryptography : 41.0.1
Platform : Linux-3.10.0-1160.81.1.el7.x86_64-x86_64-with-glibc2.2.5
```
### Additional context
None.
| 0easy
|
Title: Choosing to cancel on list menu dialog
Body: ### Has this issue already been reported?
- [X] I have searched through the existing issues.
### Is this a question rather than an issue?
- [X] This is not a question.
### What type of issue is this?
Bug
### Which Linux distribution did you use?
Manjaro/KDE
### Which AutoKey GUI did you use?
Qt
### Which AutoKey version did you use?
0.95.10-4
### How did you install AutoKey?
Arch User Repository
### Can you briefly describe the issue?
When using the dialog.list_menu in a script, choosing to cancel causes an error. I haven't tried it on many different configs but it causes the error consistently
### Can the issue be reproduced?
Always
### What are the steps to reproduce the issue?
1. Use the "List Menu" example script
2. Trigger the script
3. Press cancel or Esc
### What should have happened?
I believe it should have output None or an empty string as the choice output
### What actually happened?
The script raises a ValueError and stops
### Do you have screenshots?
_No response_
### Can you provide the output of the AutoKey command?
```bash
Script name: 'List Menu'
Traceback (most recent call last):
File "/usr/lib/python3.10/site-packages/autokey/service.py", line 485, in execute
exec(script.code, scope)
File "<string>", line 3, in <module>
File "/usr/lib/python3.10/site-packages/autokey/scripting.py", line 347, in list_menu
choice = options[int(result)]
ValueError: invalid literal for int() with base 10: ''
```
### Anything else?
There is an easy solution. Looking at the stack trace, the problem is rooted in the scripting.py file. On line 345:
`return_code, result = self._run_kdialog(title, ["--radiolist", message] + choices, kwargs)
`
when the user cancels the result is an empty string. However the next line
`choice = options[int(result)]`
does not take that into account and immediately tries to cast to an int which causes the ValueError. The self._run_kdialog function returns a nonzero return_code and an empty string for the result. If you check the return_code for a nonzero value you can just assign a None value (or whatever) to the choice and make the user deal with that choice.
```
- choice = options[int(result)]
+ if (return_code == 0):
+ choice = options[int(result)]
+ else:
+ choice = None
@@ -345 +346,348 @@
```
TL;DR never checked for an invalid value, now you should check and deal with it. | 0easy
|
Title: After inserting a fixture, update the primary key sequences
Body: After inserting a fixture:
https://github.com/piccolo-orm/piccolo/blob/8ec9d10313c9ac44916f01e27c9ef58eb22ae0d9/piccolo/apps/fixtures/commands/load.py#L57-L62
For each primary key column we should run something like:
```sql
SELECT setval('my_table_id_seq', (SELECT MAX(id) FROM my_table));
```
To make sure subsequent inserts don't fail due to unique constraint errors.
| 0easy
|
Title: BUG: Cannot read GDB file using a geometry filter (mask)
Body: - [X] I have checked that this issue has not already been reported.
- [X] I have confirmed this bug exists on the latest version of geopandas.
- [ ] (optional) I have confirmed this bug exists on the main branch of geopandas.
---
**Note**: Please read [this guide](https://matthewrocklin.com/minimal-bug-reports) detailing how to provide the necessary information for us to reproduce your bug.
#### Code Sample, a copy-pastable example
```python
gdf = gpd.read_file(
example.gdb,
driver='FileGDB',
layer='example_layer',
mask=mask_gdf,
```
#### Problem description
When running the above code using Geopandas v1.0.1, I got the following error: `ValueError: Must pass either crs or epsg.`.
The code above, however, works with v0.14.4
#### Expected Output
#### Output of ``geopandas.show_versions()``
<details>
SYSTEM INFO
-----------
python : 3.10.12 (main, Sep 11 2024, 15:47:36) [GCC 11.4.0]
executable : /usr/bin/python3
machine : Linux-6.1.85+-x86_64-with-glibc2.35
GEOS, GDAL, PROJ INFO
---------------------
GEOS : 3.11.4
GEOS lib : None
GDAL : 3.9.1
GDAL data dir: /usr/local/lib/python3.10/dist-packages/pyogrio/gdal_data/
PROJ : 9.4.1
PROJ data dir: /usr/local/lib/python3.10/dist-packages/pyproj/proj_dir/share/proj
PYTHON DEPENDENCIES
-------------------
geopandas : 1.0.1
numpy : 1.26.4
pandas : 2.2.2
pyproj : 3.7.0
shapely : 2.0.6
pyogrio : 0.10.0
geoalchemy2: None
geopy : 2.4.1
matplotlib : 3.7.1
mapclassify: None
fiona : 1.10.1
psycopg : None
psycopg2 : 2.9.9 (dt dec pq3 ext lo64)
pyarrow : 16.1.0
</details>
| 0easy
|
Title: Migrate away from deprecated github.com/hpcloud/tail
Body: /kind feature
**Describe the solution you'd like**
As https://github.com/hpcloud/tail hasn't been touched for 6 years (it is using Go v1.5.1), and the https://github.com/hpcloud organisation doesn't have any (public) members, we should really stop depending on their code. Multiple forks has been created, but it seems like https://github.com/nxadm/tail is the most popular one, so this could be a potential candidate
**Anything else you would like to add:**
The reason I found this usage of a deprecated package, is that I hit https://github.com/kubeflow/katib/issues/1769 and tried to investigate the issue
---
<!-- Don't delete this message to encourage users to support your issue! -->
Love this feature? Give it a ๐ We prioritize the features with the most ๐
| 0easy
|
Title: [CropAnnotator] - mark objects with zoomed in crop
Body: ### Description
Create `CropAnnotator` - whose task is to mark objects related to [`sv.Detections`](https://supervision.roboflow.com/detection/core/#detections) with zoomed-in crops.
Below is an example found on the internet. Of course, as part of this task, we would only add the part responsible for the crop.

The API should also allow for the selection of the location of the enlarged crop using [`sv.Position`](https://supervision.roboflow.com/geometry/core/#position) (above, below, ...), as well as the choice of the `zoom_factor`.
### API
```python
class CropAnnotator(BaseAnnotator):
def __init__(
self,
position: Position = Position.TOP_CENTER,
zoom_factor: int = 2
):
pass
def annotate(
self,
scene: np.ndarray,
detections: Detections,
) -> np.ndarray:
pass
```
### Additional
- Note: Please share a Google Colab with minimal code to test the new feature. We know it's additional work, but it will definitely speed up the review process. Each change must be tested by the reviewer. Setting up a local environment to do this is time-consuming. Please ensure that Google Colab can be accessed without any issues (make it public). Thank you! ๐๐ป
| 0easy
|
Title: Docs: Warn about using `TYPE_CHECKING` for injected types
Body: ### Summary
Typically, annotation-only imports may be skipped during runtime (for perf gains) with `if TYPE_CHECKING` blocks.
There are even linters to enforce this, ie. nag when an annotation-only import is done without `TYPE_CHECKING` block. There's https://github.com/snok/flake8-type-checking and a matching [ruff rule](https://docs.astral.sh/ruff/rules/typing-only-first-party-import/).
Moving the import to `if TYPE_CHECKING` block will raise error, as Litestar DI needs the type information:
```py
# app.py
from litestar import Litestar, get
from litestar.di import Provide
from typing import TYPE_CHECKING
if TYPE_CHECKING:
from service import Service
@get()
async def hello(service: Service) -> None:
pass
app = Litestar(
route_handlers=[hello],
dependencies={"service": Provide(Service, sync_to_thread=False)}
)
# service.py
class Service:
pass
```
```console
โฏ python app.py
Traceback (most recent call last):
File "/home/musttu/Projects/api/api/app.py", line 18, in <module>
dependencies={"service": Provide(Service, sync_to_thread=False)}
NameError: name 'Service' is not defined
```
I didn't see any mention of this caveat in the documentation - should be added.
---
Also, is there some trick to configure the linters to _not_ nag in these cases (Litestar DI), but do in others?
It seems that `flake8-type-checking` [supports FastAPI](https://github.com/snok/flake8-type-checking/issues/52) but [apparently not Litestar](https://github.com/snok/flake8-type-checking/issues?q=is%3Aissue+is%3Aopen+litestar). `ruff`probably doesn't have any special tricks for either(?) | 0easy
|
Title: Check support for Python 3.6
Body: With Python 3.6 out, we should make sure that it is properly supported by factory_boy.
This should work out of the box; but some steps are required to check it:
1. Add Python 3.6 to https://github.com/FactoryBoy/factory_boy/blob/master/tox.ini
2. Add that build in https://github.com/FactoryBoy/factory_boy/blob/master/.travis.yml
3. Make a pull request, wait for the travis build, and check that it's still working.
4. Update ``README.rst`` and classifiers at https://github.com/FactoryBoy/factory_boy/blob/master/setup.py#L57 to state that Python 3.6 is supported.
| 0easy
|
Title: [BUG] 500 when invalid overwrite url is filled in
Body: <!--
Please fill in each section below, otherwise, your issue will be closed.
This info allows django CMS maintainers to diagnose (and fix!) your issue
as quickly as possible.
-->
## Description
From time to time some of our clients manage to put random (invalid) strings as slug in the overwrite url (in advanced settings) which results in 500s.
<!--
If this is a security issue stop immediately and follow the instructions at:
http://docs.django-cms.org/en/latest/contributing/development-policies.html#reporting-security-issues
-->
## Steps to reproduce
<!--
Clear steps describing how to reproduce the issue.
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
-->
1. Go to Advanced settings
2. Enter an overwrite url containing an invalid character (for instance a space)
## Expected behaviour
A validation error is raised that says it is an invalid url.
<!--
A clear and concise description of what you expected to happen.
-->
## Actual behaviour
The cms starts raising errors.:
```
NoReverseMatch: Reverse for 'pages-details-by-slug' with keyword arguments '{'slug': 'this is an invalid slug'}' not found. 1 pattern(s) tried: ['(?P<slug>[0-9A-Za-z-_.//]+)/$']
```
<!--
A clear and concise description of what is actually happening.
-->
## Screenshots
<!--If applicable, add screenshots to help explain your problem.
-->
## Additional information (CMS/Python/Django versions)
<!--
Add any other context about the problem such as environment,
CMS/Python/Django versions, logs etc. here.
-->
## Do you want to help fix this issue?
<!--
The django CMS project is managed and kept alive by its open source community and is backed by the [django CMS Association](https://www.django-cms.org/en/about-us/). We therefore welcome any help and are grateful if people contribute to the project. Please use 'x' to check the items below.
-->
* [x] Yes, I want to help fix this issue and I will join #workgroup-pr-review on [Slack](https://www.django-cms.org/slack) to confirm with the community that a PR is welcome.
* [ ] No, I only want to report the issue.
| 0easy
|
Title: Feature: use subscribe in KafkaBroker to create handlers with explicit options propagation
Body: https://aiokafka.readthedocs.io/en/stable/consumer.html?highlight=subscribe#topic-subscription-by-pattern | 0easy
|
Title: Enhancement: Abstraction for injecting template globals
Body: ### Summary
It would be nice to have a nice abstraction for injecting globals into templates so that we don't have to do such a dance around this current process (see #basic-examples)
### Basic Example
Some examples of what is currently required I think:
@Alc-Alc's
```py
"""Template config."""
from __future__ import annotations
import os
from typing import Any
from jinja2 import Environment
from litestar import Litestar
from litestar.contrib.jinja import JinjaTemplateEngine
from litestar.template.config import TemplateConfig
def create_environment() -> Environment:
env = Environment()
env.globals['BASE_PATH'] = "/prod-path" if os.getenv("ENVIRONMENT", "dev") == "prod" else "/"
return env
config = TemplateConfig(
directory="/path/to/templates",
engine=JinjaTemplateEngine.from_environment(create_environment()),
)
app = Litestar(template_config=config)
```
or mine
```py
"""Template config."""
from __future__ import annotations
import os
from typing import Any
from jinja2 import Environment
from litestar import Litestar
from litestar.contrib.jinja import JinjaTemplateEngine
from litestar.template.config import TemplateConfig
def set_base_path(engine_instance: Any) -> None:
"""Set the base path for the template engine."""
env = os.getenv("ENVIRONMENT", "dev")
if hasattr(engine_instance, "engine") and isinstance(engine_instance.engine, Environment):
engine_instance.engine.globals["BASE_PATH"] = "/prod-path" if env == "prod" else "/"
config = TemplateConfig(
directory="/path/to/templates",
engine=JinjaTemplateEngine,
engine_callback=set_base_path,
)
app = Litestar(
template_config=config,
)
```
...and then:
```html
<a href="{{ BASE_PATH }}">some link</a>
```
### Drawbacks and Impact
- maintenance burden
### Unresolved questions
_No response_ | 0easy
|
Title: Change Requests Declined metric API
Body: The canonical definition is here: https://chaoss.community/?p=3588 | 0easy
|
Title: Improve parsing of DateTimeWidget
Body: I think the following code could be improved, by adding a break after `strptime` has returned without an exception:
https://github.com/django-import-export/django-import-export/blob/029396ddeb2c02e651f3f2b6cc666fe0eb8a647a/import_export/widgets.py#L290-L294
The current version always iterates through all patterns (by default [these](https://docs.djangoproject.com/en/5.0/ref/settings/#datetime-input-formats)), even if for example the first one already matched.
So the changed `clean()` would look like this:
```python
def clean(self, value, row=None, **kwargs):
dt = None
if not value:
return None
if isinstance(value, datetime):
dt = value
else:
for format_ in self.formats:
try:
dt = datetime.strptime(value, format_)
break # CHANGE
except (ValueError, TypeError):
continue
if dt:
if settings.USE_TZ and timezone.is_naive(dt):
dt = timezone.make_aware(dt)
return dt
raise ValueError("Enter a valid date/time.")
```
This would the be analog to the processing of the `DateWidget`, which also returns when `strptime` executes successfully:
https://github.com/django-import-export/django-import-export/blob/029396ddeb2c02e651f3f2b6cc666fe0eb8a647a/import_export/widgets.py#L244-L248 | 0easy
|
Title: Make the WHILE loop condition optional
Body: As discussed in issue #4562, if no condition is set on the WHILE loop, the condition should be `True` by default.
Here is an example :
```
*** Test Cases ***
Optional WHILE loop condition
WHILE
Log Test
END
``` | 0easy
|
Title: [tech debt] Update params for `RandomResizedCrop`
Body: Match signature of `Albumentations.RandomResizedCrop` and `TorchVision.RandomResizedCrop`
i.e. add parameter `size = [height, width]` and throw deprecated warning for `height / width` | 0easy
|
Title: Feature Branch workflow deployment to `pypi`
Body: Setup github workflow to deploy feature branch to `pypi`.
This is a dependency for issue #757. | 0easy
|
Title: Add suite and test `id` to JSON result model
Body: Model objects have an `id` attribute that is automatically generated so that the root suite has id `s1`, its first test has id `s1-t1`, its first keyword has `s1-t1-k1`, and so on. Because ids change if there are changes to data, or if different tests are selected to be run, they cannot be used for identifying suites, tests or keywords between runs, but withing a single run they can be used for pinpointing a certain item. This is used, for example, when linking to tests in the log file from the report file.
In the XML result model (i.e. `output.xml`) suites and tests have an `id` attribute. For consistency also the JSON result model should have it, especially because we are planning to support JSON results also as part of execution (#3423). `id` should be written to JSON, but Robot itself should ignore it when reading JSON. | 0easy
|
Title: Expand ~ in paths
Body: From #2820 (@tomaarsen)
Expand ~ in paths used in environment variables to the home path of the user's operating system. This is as simple as wrapping a path with path = os.path.expanduser(path). This probably ought to be done in some functions in internals.py. However, files on linux are allowed to have ~ in them. In some of those cases, you wouldn't want to expand the ~. | 0easy
|
Title: function generate_data
Body: I'm using latest pyod version on pypi. How to generate simulated data where x-axis is time? Thank you. | 0easy
|
Title: Configurable link back to Notebook Templates git repo
Body: The URL in the "execute a notebook" sidebar should be configurable, so that we can link back to the repo if users so choose.
If the configuration is called something like GIT_REPO_BASE_URL, then we can also extrapolate the URL for the individual templates to link directly back to their source code in GitHub/BitBucket (only if GIT_REPO_BASE_URL has been defined). | 0easy
|
Title: Add more unit tests to improve code coverage
Body: You can check codecov runs for the lastest master branch and find lines of code that aren't currently covered by any unit test. It would be very helpful if you can add a unit test to cover such lines of code.
Important: any new unit tests you add must run very quickly! Ideally in under 1sec. If your tests run on some data, try to use the smallest dataset possible for the test to still be useful. | 0easy
|
Title: Pagination not taking the search query into account
Body: ### Checklist
- [X] The bug is reproducible against the latest release or `master`.
- [X] There are no similar issues or pull requests to fix it yet.
### Describe the bug
When on a list page, if you preform a search query, the number of results and pages does not change according to the actual results returned by the query.
### Steps to reproduce the bug
1. Open a list page
2. Search for something that narrows down results
### Expected behavior
The result count at the bottom of the page and the number of pages should change to reflect the new results.
### Actual behavior
The result count never changes, and the number of pages stays the same, allowing the user to change pages. Once a user changes to a page outside the range of results, an empty page is shown.
### Debugging material
Here is an example, of a query that returns only 4 results, but the page still lists 15.
<img width="943" alt="Screenshot 2023-05-25 at 11 17 53" src="https://github.com/aminalaee/sqladmin/assets/44203439/eb662d13-cc07-4726-9448-bbcb73268437">
### Environment
- macOS 12.3 / Python 3.9.14 / 0.11.0
### Additional context
The bug is happening because the `ModelView.count()` method takes the whole collection into account instead of taking the search query statement as a parameter. Making the `ModelView.count_query` attribute a method that takes the statement returned by `ModelView.search_query()` as a parameter could be a way to fix this bug. | 0easy
|
Title: Docs: Code block line length
Body: ### Summary
For documentation, and only documentation, if you have an overly long code block it enters scrollable window.
We should set `blacken-docs` configuration (and eventually `ruff` when https://github.com/astral-sh/ruff/issues/8237 happens) to line lengths somewhere on the lower end (maybe 80?); this goes together with manually ensuring `.. code-block::` directives are not overly long | 0easy
|
Title: [Feature request] Add apply_to_images to Normalize
Body: | 0easy
|
Title: ไธบไปไนๅ ่ฝฝๅฎ้ข่ฎญ็ปๆจกๅไนๅๅฐฑๅกไฝไธๅจไบ๏ผๆๅคงไฝฌๅฏไปฅ่งฃ็ญไธไธๅ
Body: | 0easy
|
Title: FAB auth manager doesn't show error messages on failed log in
Body: ### Apache Airflow version
3.0.0
### What happened?
When you end up on`http://127.0.0.1:9091/auth/login/` (I think this is the FAB auth manager| Maybe it'sand enter gibberish you just get redirected to the same page with no indication of an error.
| 0easy
|
Title: docs: inline docs/includes to pages itself
Body: We have a lot of `includes/` - https://github.com/airtai/faststream/tree/main/docs/includes files created to share them between different translations. But now, we don't plan to support multilanguage documentation anymore. So, we should inline such files to usages place to make documentation sources much easy to understand, support and work with.
This is a huge work, so you can take care of any number of files you want โ all PRs are welcome!
PLEASE, DO NOT INLINE FILES ALREADY USING MULTIPLE TIMES | 0easy
|
Title: JSON.parse: unexpected character at line 1 column 1 of the JSON data
Body: I got this error last night:

The first character in the word I posted was ืข. (Language: Hebrew)
I am wondering if this is related to the language being read right-to-left? I brought this up in #301 .
After I got the error, the website was inaccessible with an error 502, verified by https://downforeveryoneorjustme.com/ . Not sure if this was just a coincidence, or if this was a terminating error. | 0easy
|
Title: refactor `.get` accessor to be generated by metaclass
Body: This is not very visible to the user right now. It's not in the docs. The only reason someone could find it is by using tab completion in IPython or viewing the `dir(.)`.
The best solution is probably make a metaclass wrapper like for `on_*` and `do_*` | 0easy
|
Title: How to load 3.8.3 gensim word2vec model by new version 4.2.0 gensim model
Body: <!--
**IMPORTANT**:
- Use the [Gensim mailing list](https://groups.google.com/forum/#!forum/gensim) to ask general or usage questions. Github issues are only for bug reports.
- Check [Recipes&FAQ](https://github.com/RaRe-Technologies/gensim/wiki/Recipes-&-FAQ) first for common answers.
Github bug reports that do not include relevant information and context will be closed without an answer. Thanks!
-->
## Problem description
Hi! I have been using gensim 3.8.1 and want to use the latest version, gensim 4.2.0.
I have been able to load models made with older gensim versions, but when I changed to the latest version, I can no longer load them.
This is the error code.
```zsh
Exception has occurred: AttributeError
Can't get attribute 'Vocab' on <module 'gensim.models.word2vec' from '/PATH'>
```
I would like to know if there is some way to load older models.
## Versions
python 3.8.9
gensim 4.2.0 | 0easy
|
Title: Add table of content to Config page documentation
Body: Add a table of contents at the top of the [config page](https://lux-api.readthedocs.io/en/latest/source/reference/config.html) in the documentation. | 0easy
|
Title: Add DayOfYear primitive
Body: - This primitive determines the day of the year a datetime column falls into
- https://pandas.pydata.org/docs/reference/api/pandas.Timestamp.dayofyear.html | 0easy
|
Title: [BUG] df.sort_values() failed when input dataframe is empty
Body: <!--
Thank you for your contribution!
Please review https://github.com/mars-project/mars/blob/master/CONTRIBUTING.rst before opening an issue.
-->
**Describe the bug**
df.sort_values() failed when input dataframe is empty.
**To Reproduce**
To help us reproducing this bug, please provide information below:
1. Your Python version
2. The version of Mars you use
3. Versions of crucial packages, such as numpy, scipy and pandas
4. Full stack of the error.
5. Minimized code to reproduce the error.
```
In [13]: import mars.tensor as mt
In [14]: import mars.dataframe as md
In [15]: df = md.DataFrame(mt.random.rand(4, 3, chunk_size=3))
In [16]: df = df[df[0] > 1].execute()
100%|โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ| 100.0/100 [00:00<00:00, 217.32it/s]
In [17]: df.sort_values(by=2).execute()
0%| | 0/100 [00:00<?, ?it/s]Unexpected error happens in <function TaskProcessor.get_next_stage_processor at 0x7f9487bc6430>
Traceback (most recent call last):
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/processor.py", line 57, in inner
return await func(processor, *args, **kwargs)
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/processor.py", line 336, in get_next_stage_processor
chunk_graph = await self._get_next_chunk_graph(self._chunk_graph_iter)
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/processor.py", line 266, in _get_next_chunk_graph
chunk_graph = await fut
File "/Users/qinxuye/Workspace/mars/mars/lib/aio/_threads.py", line 36, in to_thread
return await loop.run_in_executor(None, func_call)
File "/Users/qinxuye/miniconda3/envs/mars3.8/lib/python3.8/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/processor.py", line 261, in next_chunk_graph
return next(chunk_graph_iter)
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/preprocessor.py", line 158, in tile
for chunk_graph in chunk_graph_builder.build():
File "/Users/qinxuye/Workspace/mars/mars/core/graph/builder/chunk.py", line 272, in build
yield from self._build()
File "/Users/qinxuye/Workspace/mars/mars/core/graph/builder/chunk.py", line 266, in _build
graph = next(tile_iterator)
File "/Users/qinxuye/Workspace/mars/mars/services/task/supervisor/preprocessor.py", line 75, in __iter__
to_update_tileables = self._iter()
File "/Users/qinxuye/Workspace/mars/mars/core/graph/builder/chunk.py", line 198, in _iter
self._tile(
File "/Users/qinxuye/Workspace/mars/mars/core/graph/builder/chunk.py", line 113, in _tile
need_process = next(tile_handler)
File "/Users/qinxuye/Workspace/mars/mars/core/graph/builder/chunk.py", line 84, in _tile_handler
tiled_tileables = yield from handler.tile(tiled_tileables)
File "/Users/qinxuye/Workspace/mars/mars/core/entity/tileables.py", line 79, in tile
tiled_result = yield from tile_handler(op)
File "/Users/qinxuye/Workspace/mars/mars/dataframe/sort/core.py", line 181, in tile
return (yield from cls._tile(op))
File "/Users/qinxuye/Workspace/mars/mars/dataframe/sort/sort_values.py", line 98, in _tile
return (yield from cls._tile_dataframe(op))
File "/Users/qinxuye/Workspace/mars/mars/dataframe/sort/sort_values.py", line 66, in _tile_dataframe
return (yield from cls._tile_psrs(op, df))
File "/Users/qinxuye/Workspace/mars/mars/dataframe/sort/psrs.py", line 247, in _tile_psrs
in_df, axis_chunk_shape, _, _ = yield from cls.preprocess(op, in_data=in_data)
File "/Users/qinxuye/Workspace/mars/mars/tensor/base/psrs.py", line 63, in preprocess
chunk_size = int(axis_shape / axis_chunk_shape)
ZeroDivisionError: division by zero
0%| | 0/100 [00:00<?, ?it/s]
---------------------------------------------------------------------------
ZeroDivisionError Traceback (most recent call last)
<ipython-input-17-2ab36a920a9b> in <module>
----> 1 df.sort_values(by=2).execute()
~/Workspace/mars/mars/core/entity/tileables.py in execute(self, session, **kw)
460
461 def execute(self, session=None, **kw):
--> 462 result = self.data.execute(session=session, **kw)
463 if isinstance(result, TILEABLE_TYPE):
464 return self
~/Workspace/mars/mars/core/entity/executable.py in execute(self, session, **kw)
96
97 session = _get_session(self, session)
---> 98 return execute(self, session=session, **kw)
99
100 def _check_session(self, session: SessionType, action: str):
~/Workspace/mars/mars/deploy/oscar/session.py in execute(tileable, session, wait, new_session_kwargs, show_progress, progress_update_interval, *tileables, **kwargs)
1756 session = get_default_or_create(**(new_session_kwargs or dict()))
1757 session = _ensure_sync(session)
-> 1758 return session.execute(
1759 tileable,
1760 *tileables,
~/Workspace/mars/mars/deploy/oscar/session.py in execute(self, tileable, show_progress, *tileables, **kwargs)
1556 fut = asyncio.run_coroutine_threadsafe(coro, self._loop)
1557 try:
-> 1558 execution_info: ExecutionInfo = fut.result(
1559 timeout=self._isolated_session.timeout
1560 )
~/miniconda3/envs/mars3.8/lib/python3.8/concurrent/futures/_base.py in result(self, timeout)
437 raise CancelledError()
438 elif self._state == FINISHED:
--> 439 return self.__get_result()
440 else:
441 raise TimeoutError()
~/miniconda3/envs/mars3.8/lib/python3.8/concurrent/futures/_base.py in __get_result(self)
386 def __get_result(self):
387 if self._exception:
--> 388 raise self._exception
389 else:
390 return self._result
~/Workspace/mars/mars/deploy/oscar/session.py in _execute(session, wait, show_progress, progress_update_interval, cancelled, *tileables, **kwargs)
1707 while not cancelled.is_set():
1708 try:
-> 1709 await asyncio.wait_for(
1710 asyncio.shield(execution_info), progress_update_interval
1711 )
~/miniconda3/envs/mars3.8/lib/python3.8/asyncio/tasks.py in wait_for(fut, timeout, loop)
481
482 if fut.done():
--> 483 return fut.result()
484 else:
485 fut.remove_done_callback(cb)
~/miniconda3/envs/mars3.8/lib/python3.8/asyncio/tasks.py in _wrap_awaitable(awaitable)
682 that will later be wrapped in a Task by ensure_future().
683 """
--> 684 return (yield from awaitable.__await__())
685
686 _wrap_awaitable._is_coroutine = _is_coroutine
~/Workspace/mars/mars/deploy/oscar/session.py in wait()
100
101 async def wait():
--> 102 return await self._aio_task
103
104 self._future_local.future = fut = asyncio.run_coroutine_threadsafe(
~/Workspace/mars/mars/deploy/oscar/session.py in _run_in_background(self, tileables, task_id, progress, profiling)
894 )
895 if task_result.error:
--> 896 raise task_result.error.with_traceback(task_result.traceback)
897 if cancelled:
898 return
~/Workspace/mars/mars/services/task/supervisor/processor.py in inner(processor, *args, **kwargs)
55 async def inner(processor: "TaskProcessor", *args, **kwargs):
56 try:
---> 57 return await func(processor, *args, **kwargs)
58 except: # noqa: E722 # nosec # pylint: disable=bare-except # pragma: no cover
59 if log_when_error:
~/Workspace/mars/mars/services/task/supervisor/processor.py in get_next_stage_processor(self)
334
335 with Timer() as timer:
--> 336 chunk_graph = await self._get_next_chunk_graph(self._chunk_graph_iter)
337 if chunk_graph is None:
338 # tile finished
~/Workspace/mars/mars/services/task/supervisor/processor.py in _get_next_chunk_graph(self, chunk_graph_iter)
264
265 fut = asyncio.to_thread(next_chunk_graph)
--> 266 chunk_graph = await fut
267 return chunk_graph
268
~/Workspace/mars/mars/lib/aio/_threads.py in to_thread(func, *args, **kwargs)
34 ctx = contextvars.copy_context()
35 func_call = functools.partial(ctx.run, func, *args, **kwargs)
---> 36 return await loop.run_in_executor(None, func_call)
~/miniconda3/envs/mars3.8/lib/python3.8/concurrent/futures/thread.py in run(self)
55
56 try:
---> 57 result = self.fn(*self.args, **self.kwargs)
58 except BaseException as exc:
59 self.future.set_exception(exc)
~/Workspace/mars/mars/services/task/supervisor/processor.py in next_chunk_graph()
259 def next_chunk_graph():
260 try:
--> 261 return next(chunk_graph_iter)
262 except StopIteration:
263 return
~/Workspace/mars/mars/services/task/supervisor/preprocessor.py in tile(self, tileable_graph)
156 optimize = self._config.optimize_chunk_graph
157 meta_updated = set()
--> 158 for chunk_graph in chunk_graph_builder.build():
159 # optimize chunk graph
160 if optimize:
~/Workspace/mars/mars/core/graph/builder/chunk.py in build(self)
270
271 def build(self) -> Generator[Union[TileableGraph, ChunkGraph], None, None]:
--> 272 yield from self._build()
~/Workspace/mars/mars/core/graph/builder/chunk.py in _build(self)
264 try:
265 with enter_mode(build=True, kernel=True):
--> 266 graph = next(tile_iterator)
267 yield graph
268 except StopIteration:
~/Workspace/mars/mars/services/task/supervisor/preprocessor.py in __iter__(self)
73 def __iter__(self):
74 while self._tileable_handlers:
---> 75 to_update_tileables = self._iter()
76 if not self.cancelled:
77 yield self._cur_chunk_graph
~/Workspace/mars/mars/core/graph/builder/chunk.py in _iter(self)
196 next_tileable_handlers
197 ):
--> 198 self._tile(
199 chunk_graph,
200 tileable,
~/Workspace/mars/mars/core/graph/builder/chunk.py in _tile(self, chunk_graph, tileable, tile_handler, next_tileable_handlers, to_update_tileables, visited)
111 ):
112 try:
--> 113 need_process = next(tile_handler)
114 chunks = []
115 if need_process is not None:
~/Workspace/mars/mars/core/graph/builder/chunk.py in _tile_handler(self, tileable)
82 tiled_tileables = [self._get_data(t) for t in tiled_tileables]
83 # start to tile
---> 84 tiled_tileables = yield from handler.tile(tiled_tileables)
85 return tiled_tileables
86
~/Workspace/mars/mars/core/entity/tileables.py in tile(cls, tileables)
77 # they will be put into ChunkGraph and executed first.
78 # After execution, resume from the yield place.
---> 79 tiled_result = yield from tile_handler(op)
80 else:
81 # without iterative tiling
~/Workspace/mars/mars/dataframe/sort/core.py in tile(cls, op)
179 return (yield from cls._tile_head(op))
180 else:
--> 181 return (yield from cls._tile(op))
~/Workspace/mars/mars/dataframe/sort/sort_values.py in _tile(cls, op)
96 def _tile(cls, op):
97 if op.inputs[0].ndim == 2:
---> 98 return (yield from cls._tile_dataframe(op))
99 else:
100 return (yield from cls._tile_series(op))
~/Workspace/mars/mars/dataframe/sort/sort_values.py in _tile_dataframe(cls, op)
64 raise NotImplementedError("Only support puts NaNs at the end.")
65 # use parallel sorting by regular sampling
---> 66 return (yield from cls._tile_psrs(op, df))
67
68 @classmethod
~/Workspace/mars/mars/dataframe/sort/psrs.py in _tile_psrs(cls, op, in_data)
245 def _tile_psrs(cls, op, in_data):
246 out = op.outputs[0]
--> 247 in_df, axis_chunk_shape, _, _ = yield from cls.preprocess(op, in_data=in_data)
248
249 # stage 1: local sort and regular samples collected
~/Workspace/mars/mars/tensor/base/psrs.py in preprocess(cls, op, in_data)
61 ):
62 yield
---> 63 chunk_size = int(axis_shape / axis_chunk_shape)
64 chunk_sizes = [chunk_size for _ in range(int(axis_shape // chunk_size))]
65 if axis_shape % chunk_size > 0:
ZeroDivisionError: division by zero
```
| 0easy
|
Title: Regression: specifying workdir is treated like root
Body: In https://github.com/pypa/twine/issues/684#issuecomment-1350361116, we learned that since tox 4, invoking tox from a tox environment fails to create the environment. Here's a minimal reproducer:
```
draft $ cat > tox.ini
[testenv:parent]
deps=tox
allowlist_externals=ls
commands=
python -m tox -e child --notest
ls .tox/child
[testenv:child]
draft $ tox -e parent
parent: commands[0]> python -m tox -e child
child: OK (0.02 seconds)
congratulations :) (0.05 seconds)
parent: commands[1]> ls .tox/child
ls: .tox/child: No such file or directory
parent: exit 1 (0.01 seconds) /Users/jaraco/draft> ls .tox/child pid=69335
parent: FAIL code 1 (0.18=setup[0.02]+cmd[0.15,0.01] seconds)
evaluation failed :( (0.21 seconds)
```
I suspect one or more of the environment variables supplied in the parent env are affecting the invocation of tox and preventing the child env from being created. | 0easy
|
Title: Add Support for Radial Gradients
Body: Angular Gradients can sometimes be useful. CSS supports them, so why not expose them | 0easy
|
Title: [Bug]: `fontsize` in tables not working
Body: ### Bug summary
Specifying `fontsize` kwarg in `matplotlib.pyplot.table` doesn't have any effect.
### Code for reproduction
```Python
import matplotlib.pyplot as plt
import numpy as np
x = np.linspace(0, 10, 100)
y = x + 1
tableData = [['a', 1], ['b', 1]]
fig, ax = plt.subplots()
ax.plot(x, y)
t = ax.table(
cellText=tableData,
loc='top',
cellLoc='center',
fontsize=30
)
plt.show()
```
### Actual outcome

### Expected outcome
A table with bigger font size.
### Additional information
This works:
```
import matplotlib.pyplot as plt
import numpy as np
x = np.linspace(0, 10, 100)
y = x + 1
tableData = [['a', 1], ['b', 1]]
fig, ax = plt.subplots()
ax.plot(x, y)
t = ax.table(
cellText=tableData,
loc='top',
cellLoc='center'
)
t.set_fontsize(30) # <----------------
plt.show()
```

### Operating system
Windows 11
### Matplotlib Version
3.8.2
### Matplotlib Backend
QtAgg
### Python version
3.12.0
### Jupyter version
_No response_
### Installation
pip | 0easy
|
Title: Change default `FOR IN ZIP` mode from `SHORTEST` to `STRICT`
Body: RF 6.1 made it possible to configure what to do if lengths of lists iterated using `FOR IN ZIP` are different (#4682). The old default behavior to silently ignore items in longer lists (i.e. `SHORTEST` mode) was preserved, but the plan is to deprecate it in RF 7.0 (#4685). In RF 8 we can then change the default so that lengths must match (i.e. `STRICT` mode). The main motivation is avoiding false positives. See #4682 for mode details. | 0easy
|
Title: Email admins upon database connection error
Body: In #134, we handle certain database connection errors and return 503 instead of 500. This also suppresses the error emails sent to admins, albeit they are quite useful. Let's bring them back, ร la:
```
import logging
logging.getLogger('django.request').exception(message, response, request)
``` | 0easy
|
Title: Reminder: upgrade dependency versions in audio tutorial once their bugs are fixed.
Body: Current versions are there to:
- Avoid this bug in the tensorflow-io library:
https://github.com/tensorflow/io/issues/1687
- Avoid this bug in speechbrain library:
https://github.com/speechbrain/speechbrain/issues/1459
Remaining TODOs (need to wait until next version of tensorflow-io > 0.26.0 is released to test these:
- [ ] Try upgrading speechbrain, huggingface_hub, and tensorflow-io, tensorflow to latest versions and check if audio tutorial runs
- [ ] Try to get tensorflow_io to work on Mac M1/M2 | 0easy
|
Title: Improve search for alternative backends/clients
Body: Ideas based on https://pydanticlogfire.slack.com/archives/C06EDRBSAH3/p1728466950149999:
1. Searching for "opentelemetry" should return https://logfire.pydantic.dev/docs/guides/advanced/alternative-clients/ and https://logfire.pydantic.dev/docs/guides/advanced/alternative-backends/ as the first results
2. https://logfire.pydantic.dev/docs/guides/advanced/alternative-backends/ should mention the word 'server' and any other keywords likely to be searched
3. https://logfire.pydantic.dev/docs/why-logfire/opentelemetry/ should link to both those pages.
4. https://logfire.pydantic.dev/docs/why-logfire/opentelemetry/ should make the alternative clients point clearer - I think that's what "supports cross-language data integration" is trying to say but I'm not sure. | 0easy
|
Title: extend non-iid issue check in Datalab
Body: Currently Datalab's non-iid issue type is only detected based on `features`:
https://docs.cleanlab.ai/master/cleanlab/datalab/guide/issue_type_description.html#non-iid-issue
If the user does not input `features`, but does provide` pred_probs`, run this same check based on the `pred_probs` instead (just treat them as features).ย
Note that if the user did provide `features` or there was already a KNN graph constructed in Datalab, the results should be returned as they currently are, not using the `pred_probs` at all!
Reference: https://github.com/cleanlab/cleanlab/blob/master/cleanlab/datalab/internal/issue_manager/noniid.py | 0easy
|
Title: Uvicorn fails "quietly" upon a circular import error
Body: ### Discussed in https://github.com/encode/uvicorn/discussions/2034
<div type='discussions-op-text'>
<sup>Originally posted by **mikeedjones** July 8, 2023</sup>
From this [comment](https://github.com/tiangolo/fastapi/discussions/9838#discussioncomment-6391673) and discussion.
When there is a circular import in the app, uvicorn reports `ERROR: Error loading ASGI app. Could not import module "main".` without exposing the stacktrace which would be reported if you ran the code using `python main.py`. MRE below.
```python
# main.py
from bar import Bar
class Foo:
pass
async def app(scope, receive, send):
assert scope['type'] == 'http'
```
```python
#bar.py
from main import Foo
class Bar:
pass
```
Running with
```
uvicorn main:app
```
only results in
```
ERROR: Error loading ASGI app. Could not import module "main".
```
whereas running
```
python main.py
```
reports the far more useful
```
ImportError: cannot import name 'Bar' from partially initialized module 'bar' (most likely due to a circular import) (/workspaces/fastapi/bar.py)
```
Not sure how you'd start debugging this one I'm afraid - or if its even a "bug". For comparison syntax errors are reported correctly.</div> | 0easy
|
Title: Remove old images
Body: Remove:
- https://github.com/scanapi/scanapi/blob/master/images/overview.png
- https://github.com/scanapi/scanapi/blob/master/images/scanapi-report-example.png
We are not using them anymore. We should have the html report screenshot mentioned [here](https://github.com/scanapi/scanapi/issues/163) in this folder. | 0easy
|
Title: Using `sky launch -t <gpu_instance_type>` without specifying `--gpus` doesn't work
Body: Repro
```
(py310) sky launch --num-nodes 10 -t g4dn.12xlarge --use-spot --down -i0
sky.exceptions.ResourcesMismatchError: Infeasible resource demands found:
Instance type requested: g4dn.12xlarge
Accelerators for g4dn.12xlarge: {'T4': 4}
Accelerators requested: None
To fix: either only specify instance_type, or change the accelerators field to be consistent.
(py310) sky launch --num-nodes 10 -t p3.2xlarge --use-spot --down -i0 1 โต
sky.exceptions.ResourcesMismatchError: Infeasible resource demands found:
Instance type requested: p3.2xlarge
Accelerators for p3.2xlarge: {'V100': 1}
Accelerators requested: None
To fix: either only specify instance_type, or change the accelerators field to be consistent.
```
commit: 9d1803ba69583a558a3c2e73418c1b3f5590da3b
Don't remember if this is a deliberate choice we made. It's a surprising behavior to me. | 0easy
|
Title: Remove support for feed storage backends without feed_options
Body: Deprecated in 2.4.0. | 0easy
|
Title: [Detections] extend `from_transformers` with `class_names` support
Body: ### Description
Currently, Supervision supports class name extraction exclusively for Inference and Ultralytics libraries. Let's enhance [`from_transformers`](https://github.com/roboflow/supervision/blob/781a064d8aa46e3875378ab6aba1dfdad8bc636c/supervision/detection/core.py#L391) by incorporating support for Transformers. Take a look at [`from_inference`](https://github.com/roboflow/supervision/blob/781a064d8aa46e3875378ab6aba1dfdad8bc636c/supervision/detection/core.py#L449) and [`from_ultralytics`](https://github.com/roboflow/supervision/blob/781a064d8aa46e3875378ab6aba1dfdad8bc636c/supervision/detection/core.py#L177) for more reference.
### API
The code below should annotate the resulting image with class names.
```python
def from_transformers(cls, transformers_results: dict, id2label: Dict[int, str]) -> Detections:
pass
```
```python
import torch
import supervision as sv
from PIL import Image
from transformers import DetrImageProcessor, DetrForObjectDetection
processor = DetrImageProcessor.from_pretrained("facebook/detr-resnet-50")
model = DetrForObjectDetection.from_pretrained("facebook/detr-resnet-50")
image = Image.open(<PATH TO IMAGE>)
inputs = processor(images=image, return_tensors="pt")
with torch.no_grad():
outputs = model(**inputs)
width, height = image.size
target_size = torch.tensor([[height, width]])
results = processor.post_process_object_detection(
outputs=outputs, target_sizes=target_size)[0]
detections = sv.Detections.from_transformers(results, id2label=model.config.id2label)
bounding_box_annotator = sv.BoundingBoxAnnotator()
label_annotator = sv.LabelAnnotator()
annotated_image = bounding_box_annotator.annotate(
scene=image, detections=detections)
annotated_image = label_annotator.annotate(
scene=annotated_image, detections=detections)
```

### Additional
- [Transformers DETR Docs](https://huggingface.co/docs/transformers/en/model_doc/detr)
- Note: Please share a Google Colab with minimal code to test the new feature. We know it's additional work, but it will speed up the review process. The reviewer must test each change. Setting up a local environment to do this is time-consuming. Please ensure that Google Colab can be accessed without any issues (make it public). Thank you! ๐๐ป | 0easy
|
Title: Simplifying Open Source Contributions Through Operator Tiering from Dev aspect
Body: ### Search before continuing ๅ
ๆ็ดข๏ผๅ็ปง็ปญ
- [X] I have searched the Data-Juicer issues and found no similar feature requests. ๆๅทฒ็ปๆ็ดขไบ Data-Juicer ็ issue ๅ่กจไฝๆฏๆฒกๆๅ็ฐ็ฑปไผผ็ๅ่ฝ้ๆฑใ
### Description ๆ่ฟฐ
### Current state
The complete process for developing an operator can be quite demanding. This involves many things like
- adhering to coding styles
- adding new StatsKeys
- creating a new operator file, considering decorators, unified model management of HF, handling batched operations, managing paths for Mapper operators
- writing unit tests
- documenting the operator
- and ensuring it is fusible.
Although some of these steps are optional and can be automatically finished, they present additional cognitive and development burdens for new contributors, especially when compared to writing a simple demo operator from scratch.
### A potential solution
To reduce the entry barriers, I propose implementing a tiered labeling system for operators, such as `alpha_op`, `beta_op`, and `stable_op`.
- Alpha Operators: Only need to showcase new functionalities and serve as simple demos contributed by the community.
- Beta Operators: Must meet the mandatory requirements outlined in the DJ-developer guide.
- Stable Operators: In addition to satisfying beta_op criteria, they should fulfill optional recommendations in the DJ-developer guide, and write and pass both single-node and distributed unit tests.
Welcome better suggestions and more detailed implementation plans to enhance this proposal further.
### Use case ไฝฟ็จๅบๆฏ
_No response_
### Additional ้ขๅคไฟกๆฏ
_No response_
### Are you willing to submit a PR for this feature? ๆจๆฏๅฆไนๆไธบๆญคๅ่ฝๆไบคไธไธช PR๏ผ
- [X] Yes I'd like to help by submitting a PR! ๆฏ็๏ผๆๆฟๆๆไพๅธฎๅฉๅนถๆไบคไธไธชPR๏ผ | 0easy
|
Title: Improved handling of navigator in non-lazy signals
Body: With the improved handling of navigator generation for lazy signals in https://github.com/hyperspy/hyperspy/pull/2631, it would be nice to extend parts of this to non-lazy signals as well.
Currently, when `s.plot()` is used for non-lazy signals, the navigator signal is discarded after the plot is closed.
For very large datasets, the creation of the navigator signal can take a couple of seconds. So it would be nice if the navigator signal for non-lazy signals was stored, similarly as non-lazy signals. | 0easy
|
Title: DateTime: Support `datetime.date` as an input format with date related keywords
Body: Currently the library only accepts `datetime.datetime`, strings (timestamps) and numbers (epoch seconds). Supporting also `datetime.date` would be trivial and it would mean users don't need to convert `datetime.date` objects they get to a supported format themselves. Based on a discussion on our Slack, `datetime.date` objects can be produced at least by FakerLibrary. | 0easy
|
Title: Help add more datasets to our Data-Centric AI benchmark
Body: Good benchmarking is critical for the success of our tools! Our benchmarks are very unique compared to standard ML benchmarks which assess test accuracy. Instead our benchmarks assess how accurately different algorithms can identify issues in the data. Adding new datasets to our benchmarks and re-running them could make very interesting material for a blogpost!
In particular, our label error detection benchmark could greatly benefit from the addition of more datasets:
https://github.com/cleanlab/label-error-detection-benchmarks
It should be pretty evident how to run the existing benchmark code on a new image dataset once it has been added.
Some ideas for good datasets to add to this benchmark include:
* Clothing1M: https://paperswithcode.com/dataset/clothing1m
* Animal-10N: https://arxiv.org/abs/2103.07756
* Cifar-10H: https://github.com/jcpeterson/cifar-10h (as given label for each example, can take the worst label chosen across any of the annotators)
* Cats4ML: https://github.com/greg1232/dataperf-vision-creation
Our Out-of-Distribution benchmark could also benefit from the addition of more dataset pairs which are standard in the OOD research community: https://github.com/cleanlab/ood-detection-benchmarks | 0easy
|
Title: Support for additional icons
Body: ### Checklist
- [X] There are no similar issues or pull requests for this yet.
### Is your feature related to a problem? Please describe.
_No response_
### Describe the solution you would like.
eg how FastAPI-Admin can specify tabler icons https://tablericons.com/
### Describe alternatives you considered
_No response_
### Additional context
_No response_ | 0easy
|
Title: Regression in environment var inheritance in tox 4
Body: ## Issue
I'm not sure if this is a regression, or if I am using a bad pattern that worked by chance in tox3.
I use the base [testenv] to set_env some vars and run unit tests. Then I have a [testenv:functional] that inherits these env vars (most of which are still relevant) but changes the var governing where the tests are located. I use that as an "abstract" testenv to hold the setenv, commands, etc., for python-version-specific testenvs like [testenv:functional-py39], [testenv:functional-py310], etc. In tox3 this worked great, but in tox 4, the changes to set_env in the "abstract" testenv aren't picked up when I call 'tox -e functional-py39'.
Here's a ``tox.ini`` for this that works in both tox3 and tox4.
```ini
[tox]
envlist = py39,functional-py39
skipsdist=true
[testenv]
setenv =
OS_TEST_PATH=./cinderclient/tests/unit
allowlist_externals = bash
commands = bash ts.sh testenv
# abstract env
[testenv:functional]
setenv =
{[testenv]setenv}
OS_TEST_PATH=./cinderclient/tests/functional
commands = bash ts.sh functional
# python-version-specific env
[testenv:functional-py39]
setenv = {[testenv:functional]setenv}
commands = {[testenv:functional]commands}
```
All ``ts.sh`` does is echo the value of the env var of interest to a file:
```bash
#!/bin/bash
echo "$1: $OS_TEST_PATH" >> tmpfile
```
Using the ``tox.ini`` with both tox3 and tox4:
```console
โฏ ./tox4 --version >> tmpfile
โฏ ./tox4
py39: commands[0]> bash ts.sh testenv
py39: OK โ in 0.06 seconds
functional-py39: commands[0]> bash ts.sh functional
py39: OK (0.06=setup[0.06]+cmd[0.00] seconds)
functional-py39: OK (0.05=setup[0.05]+cmd[0.00] seconds)
congratulations :) (0.14 seconds)
โฏ tox --version >> tmpfile
โฏ tox
py39 create: /home/brosmait/scratch/lang/python/tox/.tox/py39
py39 run-test-pre: PYTHONHASHSEED='2285170242'
py39 run-test: commands[0] | bash ts.sh testenv
functional-py39 create: /home/brosmait/scratch/lang/python/tox/.tox/functional-py39
functional-py39 run-test-pre: PYTHONHASHSEED='2285170242'
functional-py39 run-test: commands[0] | bash ts.sh functional
____________________________________________________________ summary _____________________________________________________________
py39: commands succeeded
functional-py39: commands succeeded
congratulations :)
```
Looking at the created file:
```console
โฏ cat tmpfile
4.2.5 from /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/tox/__init__.py
testenv: ./cinderclient/tests/unit
functional: ./cinderclient/tests/unit
3.28.0 imported from /usr/lib/python3.10/site-packages/tox/__init__.py
testenv: ./cinderclient/tests/unit
functional: ./cinderclient/tests/functional
```
You can see that tox3 uses the correct value for the env var for the functional tests, but tox4 is using the original value.
## Environment
Provide at least:
- OS: Fedora release 36
- `pip list` of the host Python where `tox` is installed:
```console
โฏ pip list
Package Version
------------- -------
cachetools 5.2.0
chardet 5.1.0
colorama 0.4.6
distlib 0.3.6
filelock 3.9.0
packaging 22.0
pip 22.3.1
platformdirs 2.6.2
pluggy 1.0.0
pyproject_api 1.4.0
setuptools 65.6.3
tomli 2.0.1
tox 4.2.5
virtualenv 20.17.1
wheel 0.38.4
```
## Output of running tox
Provide the output of `tox -rvv`:
```console
โฏ ./tox4 -rvv
py39: 83 W remove tox env folder /home/brosmait/scratch/lang/python/tox/.tox/py39 [tox/tox_env/api.py:321]
py39: 103 I find interpreter for spec PythonSpec(major=3, minor=9) [virtualenv/discovery/builtin.py:56]
py39: 103 D discover exe for PythonInfo(spec=CPython3.10.9.final.0-64, exe=/home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/bin/python, platform=linux, version='3.10.9 (main, Dec 7 2022, 00:00:00) [GCC 12.2.1 20221121 (Red Hat 12.2.1-4)]', encoding_fs_io=utf-8-utf-8) in /usr [virtualenv/discovery/py_info.py:437]
py39: 103 D filesystem is case-sensitive [virtualenv/info.py:24]
py39: 103 D got python info of /usr/bin/python3.10 from /home/brosmait/.local/share/virtualenv/py_info/1/8a94588eda9d64d9e9a351ab8144e55b1fabf5113b54e67dd26a8c27df0381b3.json [virtualenv/app_data/via_disk_folder.py:129]
py39: 104 I proposed PythonInfo(spec=CPython3.10.9.final.0-64, system=/usr/bin/python3.10, exe=/home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/bin/python, platform=linux, version='3.10.9 (main, Dec 7 2022, 00:00:00) [GCC 12.2.1 20221121 (Red Hat 12.2.1-4)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
py39: 104 D discover PATH[0]=/home/brosmait/.pyenv/plugins/pyenv-virtualenv/shims [virtualenv/discovery/builtin.py:108]
py39: 104 D discover PATH[1]=/home/brosmait/.pyenv/shims [virtualenv/discovery/builtin.py:108]
py39: 104 D got python info of /home/brosmait/.pyenv/shims/python3.9 from /home/brosmait/.local/share/virtualenv/py_info/1/df5750520713fe69cbf62cb311668a6a03aa778e2b7543129d2ab17ce2716c13.json [virtualenv/app_data/via_disk_folder.py:129]
py39: 105 D got python info of /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 from /home/brosmait/.local/share/virtualenv/py_info/1/e9b605bd999ef794c335c758558757307e733b618c62dd3ca2fa6d9c79695fff.json [virtualenv/app_data/via_disk_folder.py:129]
py39: 105 I proposed PathPythonInfo(spec=CPython3.9.14.final.0-64, system=/home/brosmait/.pyenv/versions/3.9.14/bin/python3.9, exe=/home/brosmait/.pyenv/shims/python3.9, platform=linux, version='3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
py39: 105 D accepted PathPythonInfo(spec=CPython3.9.14.final.0-64, system=/home/brosmait/.pyenv/versions/3.9.14/bin/python3.9, exe=/home/brosmait/.pyenv/shims/python3.9, platform=linux, version='3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
py39: 120 I create virtual environment via CPython3Posix(dest=/home/brosmait/scratch/lang/python/tox/.tox/py39, clear=False, no_vcs_ignore=False, global=False) [virtualenv/run/session.py:48]
py39: 120 D create folder /home/brosmait/scratch/lang/python/tox/.tox/py39/bin [virtualenv/util/path/_sync.py:9]
py39: 120 D create folder /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages [virtualenv/util/path/_sync.py:9]
py39: 120 D write /home/brosmait/scratch/lang/python/tox/.tox/py39/pyvenv.cfg [virtualenv/create/pyenv_cfg.py:30]
py39: 120 D home = /home/brosmait/.pyenv/versions/3.9.14/bin [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D implementation = CPython [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D version_info = 3.9.14.final.0 [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D virtualenv = 20.17.1 [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D include-system-site-packages = false [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D base-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D base-exec-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D base-executable = /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 [virtualenv/create/pyenv_cfg.py:34]
py39: 120 D symlink /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 to /home/brosmait/scratch/lang/python/tox/.tox/py39/bin/python [virtualenv/util/path/_sync.py:28]
py39: 121 D create virtualenv import hook file /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/_virtualenv.pth [virtualenv/create/via_global_ref/api.py:89]
py39: 121 D create /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/_virtualenv.py [virtualenv/create/via_global_ref/api.py:92]
py39: 121 D ============================== target debug ============================== [virtualenv/run/session.py:50]
py39: 121 D debug via /home/brosmait/scratch/lang/python/tox/.tox/py39/bin/python /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/create/debug.py [virtualenv/create/creator.py:197]
py39: 121 D {
"sys": {
"executable": "/home/brosmait/scratch/lang/python/tox/.tox/py39/bin/python",
"_base_executable": "/home/brosmait/scratch/lang/python/tox/.tox/py39/bin/python",
"prefix": "/home/brosmait/scratch/lang/python/tox/.tox/py39",
"base_prefix": "/home/brosmait/.pyenv/versions/3.9.14",
"real_prefix": null,
"exec_prefix": "/home/brosmait/scratch/lang/python/tox/.tox/py39",
"base_exec_prefix": "/home/brosmait/.pyenv/versions/3.9.14",
"path": [
"/home/brosmait/.pyenv/versions/3.9.14/lib/python39.zip",
"/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9",
"/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/lib-dynload",
"/home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages"
],
"meta_path": [
"<class '_virtualenv._Finder'>",
"<class '_frozen_importlib.BuiltinImporter'>",
"<class '_frozen_importlib.FrozenImporter'>",
"<class '_frozen_importlib_external.PathFinder'>"
],
"fs_encoding": "utf-8",
"io_encoding": "utf-8"
},
"version": "3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]",
"makefile_filename": "/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/config-3.9-x86_64-linux-gnu/Makefile",
"os": "<module 'os' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/os.py'>",
"site": "<module 'site' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/site.py'>",
"datetime": "<module 'datetime' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/datetime.py'>",
"math": "<module 'math' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so'>",
"json": "<module 'json' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/json/__init__.py'>"
} [virtualenv/run/session.py:51]
py39: 136 I add seed packages via FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=/home/brosmait/.local/share/virtualenv) [virtualenv/run/session.py:55]
py39: 137 D install setuptools from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/setuptools-65.6.3-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
py39: 137 D install pip from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/pip-22.3.1-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
py39: 137 D install wheel from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/wheel-0.38.4-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
py39: 138 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/distutils-precedence.pth to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/distutils-precedence.pth [virtualenv/util/path/_sync.py:36]
py39: 138 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/wheel [virtualenv/util/path/_sync.py:36]
py39: 138 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/pip [virtualenv/util/path/_sync.py:36]
py39: 138 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/_distutils_hack to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/_distutils_hack [virtualenv/util/path/_sync.py:36]
py39: 139 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/pkg_resources to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/pkg_resources [virtualenv/util/path/_sync.py:36]
py39: 142 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel-0.38.4.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/wheel-0.38.4.dist-info [virtualenv/util/path/_sync.py:36]
py39: 143 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel-0.38.4.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/wheel-0.38.4.virtualenv [virtualenv/util/path/_sync.py:36]
py39: 144 D generated console scripts wheel-3.9 wheel wheel3 wheel3.9 [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
py39: 147 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/setuptools [virtualenv/util/path/_sync.py:36]
py39: 165 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools-65.6.3.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/setuptools-65.6.3.dist-info [virtualenv/util/path/_sync.py:36]
py39: 166 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools-65.6.3.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/setuptools-65.6.3.virtualenv [virtualenv/util/path/_sync.py:36]
py39: 166 D generated console scripts [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
py39: 178 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip-22.3.1.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/pip-22.3.1.dist-info [virtualenv/util/path/_sync.py:36]
py39: 178 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip-22.3.1.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/py39/lib/python3.9/site-packages/pip-22.3.1.virtualenv [virtualenv/util/path/_sync.py:36]
py39: 179 D generated console scripts pip3.9 pip pip-3.9 pip3 [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
py39: 179 I add activators for Bash, CShell, Fish, Nushell, PowerShell, Python [virtualenv/run/session.py:61]
py39: 180 D write /home/brosmait/scratch/lang/python/tox/.tox/py39/pyvenv.cfg [virtualenv/create/pyenv_cfg.py:30]
py39: 180 D home = /home/brosmait/.pyenv/versions/3.9.14/bin [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D implementation = CPython [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D version_info = 3.9.14.final.0 [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D virtualenv = 20.17.1 [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D include-system-site-packages = false [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D base-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D base-exec-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
py39: 180 D base-executable = /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 [virtualenv/create/pyenv_cfg.py:34]
py39: 181 W commands[0]> bash ts.sh testenv [tox/tox_env/api.py:427]
py39: 183 I exit 0 (0.00 seconds) /home/brosmait/scratch/lang/python/tox> bash ts.sh testenv pid=315709 [tox/execute/api.py:275]
py39: OK โ in 0.1 seconds
functional-py39: 183 W remove tox env folder /home/brosmait/scratch/lang/python/tox/.tox/functional-py39 [tox/tox_env/api.py:321]
functional-py39: 201 I find interpreter for spec PythonSpec(major=3, minor=9) [virtualenv/discovery/builtin.py:56]
functional-py39: 201 I proposed PythonInfo(spec=CPython3.10.9.final.0-64, system=/usr/bin/python3.10, exe=/home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/bin/python, platform=linux, version='3.10.9 (main, Dec 7 2022, 00:00:00) [GCC 12.2.1 20221121 (Red Hat 12.2.1-4)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
functional-py39: 201 D discover PATH[0]=/home/brosmait/.pyenv/plugins/pyenv-virtualenv/shims [virtualenv/discovery/builtin.py:108]
functional-py39: 202 D discover PATH[1]=/home/brosmait/.pyenv/shims [virtualenv/discovery/builtin.py:108]
functional-py39: 202 I proposed PathPythonInfo(spec=CPython3.9.14.final.0-64, system=/home/brosmait/.pyenv/versions/3.9.14/bin/python3.9, exe=/home/brosmait/.pyenv/shims/python3.9, platform=linux, version='3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:63]
functional-py39: 202 D accepted PathPythonInfo(spec=CPython3.9.14.final.0-64, system=/home/brosmait/.pyenv/versions/3.9.14/bin/python3.9, exe=/home/brosmait/.pyenv/shims/python3.9, platform=linux, version='3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]', encoding_fs_io=utf-8-utf-8) [virtualenv/discovery/builtin.py:65]
functional-py39: 202 I create virtual environment via CPython3Posix(dest=/home/brosmait/scratch/lang/python/tox/.tox/functional-py39, clear=False, no_vcs_ignore=False, global=False) [virtualenv/run/session.py:48]
functional-py39: 203 D create folder /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/bin [virtualenv/util/path/_sync.py:9]
functional-py39: 203 D create folder /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages [virtualenv/util/path/_sync.py:9]
functional-py39: 203 D write /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/pyvenv.cfg [virtualenv/create/pyenv_cfg.py:30]
functional-py39: 203 D home = /home/brosmait/.pyenv/versions/3.9.14/bin [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D implementation = CPython [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D version_info = 3.9.14.final.0 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D virtualenv = 20.17.1 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D include-system-site-packages = false [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D base-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D base-exec-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D base-executable = /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 203 D symlink /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/bin/python [virtualenv/util/path/_sync.py:28]
functional-py39: 203 D create virtualenv import hook file /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/_virtualenv.pth [virtualenv/create/via_global_ref/api.py:89]
functional-py39: 203 D create /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/_virtualenv.py [virtualenv/create/via_global_ref/api.py:92]
functional-py39: 203 D ============================== target debug ============================== [virtualenv/run/session.py:50]
functional-py39: 203 D debug via /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/bin/python /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/create/debug.py [virtualenv/create/creator.py:197]
functional-py39: 203 D {
"sys": {
"executable": "/home/brosmait/scratch/lang/python/tox/.tox/functional-py39/bin/python",
"_base_executable": "/home/brosmait/scratch/lang/python/tox/.tox/functional-py39/bin/python",
"prefix": "/home/brosmait/scratch/lang/python/tox/.tox/functional-py39",
"base_prefix": "/home/brosmait/.pyenv/versions/3.9.14",
"real_prefix": null,
"exec_prefix": "/home/brosmait/scratch/lang/python/tox/.tox/functional-py39",
"base_exec_prefix": "/home/brosmait/.pyenv/versions/3.9.14",
"path": [
"/home/brosmait/.pyenv/versions/3.9.14/lib/python39.zip",
"/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9",
"/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/lib-dynload",
"/home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages"
],
"meta_path": [
"<class '_virtualenv._Finder'>",
"<class '_frozen_importlib.BuiltinImporter'>",
"<class '_frozen_importlib.FrozenImporter'>",
"<class '_frozen_importlib_external.PathFinder'>"
],
"fs_encoding": "utf-8",
"io_encoding": "utf-8"
},
"version": "3.9.14 (main, Oct 4 2022, 17:06:42) \n[GCC 12.2.1 20220819 (Red Hat 12.2.1-2)]",
"makefile_filename": "/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/config-3.9-x86_64-linux-gnu/Makefile",
"os": "<module 'os' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/os.py'>",
"site": "<module 'site' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/site.py'>",
"datetime": "<module 'datetime' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/datetime.py'>",
"math": "<module 'math' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/lib-dynload/math.cpython-39-x86_64-linux-gnu.so'>",
"json": "<module 'json' from '/home/brosmait/.pyenv/versions/3.9.14/lib/python3.9/json/__init__.py'>"
} [virtualenv/run/session.py:51]
functional-py39: 219 I add seed packages via FromAppData(download=False, pip=bundle, setuptools=bundle, wheel=bundle, via=copy, app_data_dir=/home/brosmait/.local/share/virtualenv) [virtualenv/run/session.py:55]
functional-py39: 220 D install pip from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/pip-22.3.1-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
functional-py39: 220 D install setuptools from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/setuptools-65.6.3-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
functional-py39: 220 D install wheel from wheel /home/brosmait/repos/openstack/python-cinderclient/.tox/.tox/lib/python3.10/site-packages/virtualenv/seed/wheels/embed/wheel-0.38.4-py3-none-any.whl via CopyPipInstall [virtualenv/seed/embed/via_app_data/via_app_data.py:47]
functional-py39: 221 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/wheel [virtualenv/util/path/_sync.py:36]
functional-py39: 221 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/pip [virtualenv/util/path/_sync.py:36]
functional-py39: 221 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/distutils-precedence.pth to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/distutils-precedence.pth [virtualenv/util/path/_sync.py:36]
functional-py39: 222 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/_distutils_hack to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/_distutils_hack [virtualenv/util/path/_sync.py:36]
functional-py39: 222 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/pkg_resources to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/pkg_resources [virtualenv/util/path/_sync.py:36]
functional-py39: 224 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel-0.38.4.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/wheel-0.38.4.dist-info [virtualenv/util/path/_sync.py:36]
functional-py39: 226 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/wheel-0.38.4-py3-none-any/wheel-0.38.4.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/wheel-0.38.4.virtualenv [virtualenv/util/path/_sync.py:36]
functional-py39: 227 D generated console scripts wheel-3.9 wheel wheel3 wheel3.9 [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
functional-py39: 229 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/setuptools [virtualenv/util/path/_sync.py:36]
functional-py39: 249 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools-65.6.3.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/setuptools-65.6.3.dist-info [virtualenv/util/path/_sync.py:36]
functional-py39: 250 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/setuptools-65.6.3-py3-none-any/setuptools-65.6.3.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/setuptools-65.6.3.virtualenv [virtualenv/util/path/_sync.py:36]
functional-py39: 250 D generated console scripts [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
functional-py39: 261 D copy directory /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip-22.3.1.dist-info to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/pip-22.3.1.dist-info [virtualenv/util/path/_sync.py:36]
functional-py39: 262 D copy /home/brosmait/.local/share/virtualenv/wheel/3.9/image/1/CopyPipInstall/pip-22.3.1-py3-none-any/pip-22.3.1.virtualenv to /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/lib/python3.9/site-packages/pip-22.3.1.virtualenv [virtualenv/util/path/_sync.py:36]
functional-py39: 262 D generated console scripts pip3 pip3.9 pip pip-3.9 [virtualenv/seed/embed/via_app_data/pip_install/base.py:41]
functional-py39: 262 I add activators for Bash, CShell, Fish, Nushell, PowerShell, Python [virtualenv/run/session.py:61]
functional-py39: 263 D write /home/brosmait/scratch/lang/python/tox/.tox/functional-py39/pyvenv.cfg [virtualenv/create/pyenv_cfg.py:30]
functional-py39: 263 D home = /home/brosmait/.pyenv/versions/3.9.14/bin [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D implementation = CPython [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D version_info = 3.9.14.final.0 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D virtualenv = 20.17.1 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D include-system-site-packages = false [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D base-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D base-exec-prefix = /home/brosmait/.pyenv/versions/3.9.14 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 263 D base-executable = /home/brosmait/.pyenv/versions/3.9.14/bin/python3.9 [virtualenv/create/pyenv_cfg.py:34]
functional-py39: 264 W commands[0]> bash ts.sh functional [tox/tox_env/api.py:427]
functional-py39: 266 I exit 0 (0.00 seconds) /home/brosmait/scratch/lang/python/tox> bash ts.sh functional pid=315721 [tox/execute/api.py:275]
py39: OK (0.10=setup[0.10]+cmd[0.00] seconds)
functional-py39: OK (0.08=setup[0.08]+cmd[0.00] seconds)
congratulations :) (0.21 seconds)
โฏ cat tmpfile
testenv: ./cinderclient/tests/unit
functional: ./cinderclient/tests/unit
```
## Minimal example
If possible, provide a minimal reproducer for the issue:
(see file above)
## Observations
1. The "abstract" testenv is necessary, that is, the bug isn't reproducible if you just have 2 testenvs.
2. I think this may have something to do with lazy loading of environments in tox4, because the bug doesn't occur if we *use* that intermediate "abstract" environment:
```console
โฏ ./tox4 -e py39,functional,functional-py39
py39: commands[0]> bash ts.sh testenv
py39: OK โ in 0.02 seconds
functional: commands[0]> bash ts.sh functional
functional: OK โ in 0 seconds
functional-py39: commands[0]> bash ts.sh functional
py39: OK (0.02=setup[0.02]+cmd[0.00] seconds)
functional: OK (0.00=setup[0.00]+cmd[0.00] seconds)
functional-py39: OK (0.00=setup[0.00]+cmd[0.00] seconds)
congratulations :) (0.05 seconds)
โฏ cat tmpfile
testenv: ./cinderclient/tests/unit
functional: ./cinderclient/tests/functional
functional: ./cinderclient/tests/functional
```
... though, of course, the whole point of the "abstract" testenv is that you normally don't use it directly.
3. The output for 'tox c' doesn't include the change to the env var:
```console
โฏ ./tox4 c -k set_env -e py39,functional,functional-py39
[testenv:py39]
set_env =
OS_TEST_PATH=./cinderclient/tests/unit
PIP_DISABLE_PIP_VERSION_CHECK=1
PYTHONHASHSEED=4137894661
PYTHONIOENCODING=utf-8
[testenv:functional]
set_env =
OS_TEST_PATH=./cinderclient/tests/unit
PIP_DISABLE_PIP_VERSION_CHECK=1
PYTHONHASHSEED=4137894661
PYTHONIOENCODING=utf-8
[testenv:functional-py39]
set_env =
OS_TEST_PATH=./cinderclient/tests/unit
PIP_DISABLE_PIP_VERSION_CHECK=1
PYTHONHASHSEED=4137894661
PYTHONIOENCODING=utf-8
```
This seems kind of bad to me, because people not using the "abstract" testenv will still see misleading results from the 'tox c' output. In other words, if you only had [testenv] and [testenv:functional] defined (forget about the "abstract" business), 'tox c' does *not* show the value that tox is actually using for OS_TEST_PATH when you call 'tox -e functional', which makes debugging the config file difficult.
4. If the pattern I'm using to have an "abstract" job is stupid, please let me know what a better way to do this is and I'll be happy to use it. But if it's a reasonable thing to do, I imagine I'm not the only one doing it, and it's going to cause difficulties upgrading from tox3 to tox4. (I mean, the workaround is easy enough, just violate the DRY principle and explicitly set all the env vars, but you have to know to do it.) | 0easy
|
Title: [FEA] nodes/edges assign
Body: **Is your feature request related to a problem? Please describe.**
Typical node/edge equiv code for pandas `assign` is hard to read and annoying to write, cleaner w/ a native form
Today:
```python
g3 = (g2
.nodes(lambda g2: g2._nodes.assign(mcapLog10=g2._nodes['marketCap'].apply(np.log10)))
)
```
**Describe the solution you'd like**
New `nodes_assign()`, `edges_assign()` that generalize pandas `assign()`, similar to what we did for `pipe()`:
```python
g3 = (g2
.nodes_assign(
mcapLog10=g2._nodes['marketCap'].apply(np.log10)
)
.nodes_assign(lambda n2: {
'mcapLog10': n2['marketCap'].apply(np.log10) # reads _nodes df from current pipeline
})
)
```
And same for `edges_assign()`
| 0easy
|
Title: Request to add random walk index indicator
Body: Can random walk index indicator be added to the panda-ta library ? thanks
References:
[Technical Indicators](https://www.technicalindicators.net/indicators-technical-analysis/168-rwi-random-walk-index)
[trading sim](https://tradingsim.com/blog/random-walk-index/)
[linnsoft](https://www.linnsoft.com/techind/random-walk-index)
[fmlabs](https://www.fmlabs.com/reference/default.htm?url=rwi.htm)
| 0easy
|
Title: [RFC] Add option to skip reloading on validators.validate
Body: On lines
https://github.com/rochacbruno/dynaconf/blob/master/dynaconf/validator.py#L202-L205
`validators.validate` is calling `from_env` and it makes the variables to be reloaded from source files.
This is not good for testing.
Solution:
1. Check if validation is happening on the same current env, then don't reload
2. accept argument `reload=False` that will cause those lines to pass `settings` direct to the `validate_items` | 0easy
|
Title: Improve contributing documentation
Body: Things that we need to do:
- [ ] How to configure the development environment and run the tests
- [ ] Explain usage of each mode
This can also be break down into different PRs.
| 0easy
|
Title: Revert changes for local docs build once related sphinx issue is closed
Body: In MR #2367, changes were made in `docs/Makefile` to allow docs to be build locally using the `make html` command. This was needed due to errors that happened when attempting to built the docs with Featuretools installed in editable mode. Docs builds failing in editable mode *might* be related to an issue with sphinx.
When sphinx issue 10943 (https://github.com/sphinx-doc/sphinx/issues/10943) has been closed and resolved, we should revert the changes that were mode to the Makefile as indicated by the comments here:
```
.PHONY: html
html:
# Remove the following line when sphinx issue (https://github.com/sphinx-doc/sphinx/issues/10943) is closed
python -m pip install .. --quiet --no-dependencies
$(SPHINXBUILD) -b html $(ALLSPHINXOPTS) $(BUILDDIR)/html $(SPHINXOPTS)
# Remove the following line when sphinx issue (https://github.com/sphinx-doc/sphinx/issues/10943) is closed
python -m pip install -e .. --quiet --no-dependencies
@echo
@echo "Build finished. The HTML pages are in $(BUILDDIR)/html."
` | 0easy
|
Title: st2-api-key not obfuscated when using core.http?
Body: ## SUMMARY
I have 2 ST2 instances (independent of each other and in different networks) and want them to communicate with each other via API using API keys. However, when providing st2-api-key to headers of action core.http, the API key is visible in plain-text in both st2web and in CLI. This is not desirable, as I want the users to be able to use the keys, but not unintentionally share them during any screen sharing sessions. Masking is set in the config for both [api] and [log] (and any actions I've created that use the "secret" tag are masked properly) in st2.conf and I've even tried adding st2-api-key into mask_secrets_blacklist. I've tried to clone the runner, but headers are not overridable (can't just create my own http runner with headers marked as "secret"). Before going on and writing my own http as a python action, I wanted to ask whether I'm doing something wrong, as it seems obvious to me that any auth info should be obfuscated by default.
### STACKSTORM VERSION
[root@st2 st2]# st2 --version
st2 3.7.0, on Python 3.8.12
[root@st2 st2]#
##### OS, environment, install method
custom install on a RHEL8
## Steps to reproduce the problem
Use core.http with st2-api-key: <value> in st2web.
## Expected Results
Expected the value of the API key to be obfuscated.
## Actual Results
The API key is visible in plaintext.
Making sure to follow these steps will guarantee the quickest resolution possible.
Thanks!
| 0easy
|
Title: Will FastAPI support QUERY http method? "app.query"
Body: ### Discussed in https://github.com/fastapi/fastapi/discussions/6049
<div type='discussions-op-text'>
<sup>Originally posted by **FilipeMarch** December 16, 2022</sup>
### First Check
- [X] I added a very descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I searched the FastAPI documentation, with the integrated search.
- [X] I already searched in Google "How to X in FastAPI" and didn't find any information.
- [X] I already read and followed all the tutorial in the docs and didn't find an answer.
- [X] I already checked if it is not related to FastAPI but to [Pydantic](https://github.com/samuelcolvin/pydantic).
- [X] I already checked if it is not related to FastAPI but to [Swagger UI](https://github.com/swagger-api/swagger-ui).
- [X] I already checked if it is not related to FastAPI but to [ReDoc](https://github.com/Redocly/redoc).
### Commit to Help
- [X] I commit to help with one of those options ๐
### Example Code
```python
from fastapi import FastAPI
app = FastAPI()
@app.query('/query/subjects')
def query_subjects(schema: ArbitrarySchema):
with Session(engine) as db:
subjects = db.query(Subject).all()
return schema(**subjects)
# something like this
```
### Description
There is a new HTTP method called QUERY, I discovered it this week and it is super interesting!
https://www.ietf.org/archive/id/draft-ietf-httpbis-safe-method-w-body-02.html
I was trying to make a GET route that would accept a request body, but started receiving an error on Swagger UI:
[TypeError: Request has method 'GET' and cannot have a body](https://github.com/swagger-api/swagger-ui/issues/5891#issuecomment-1340591857)
The idea is that sometimes we need to make a big or complex query, and this is the scenario in which we can see the advantages of GraphQL, and although I have seen that [we can integrate FastAPI with GraphQL](https://fastapi.tiangolo.com/advanced/graphql/), I was wondering if FastAPI will ever be able to simply accept a QUERY method like, for example, `app.query("/query/subjects")`.
For example, suppose I have this object:
```yml
{
"id": 1,
"name": "Math"
"tags": [
{
"id": 1,
"name": "Algebra",
"number_of_clicks": 1,
"number_of_questions": 7,
"number_of_answers": 3,
"number_of_comments": 2,
"number_of_votes": 1,
}]
"topics": [
{
"id": 1,
"name": "Linear Equations",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_tutorials": 1,
"number_of_questions": 7,
"posts": [
{
"id": 1,
"title": "How to solve linear equations?",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_answers": 3,
"number_of_comments": 2,
"number_of_votes": 1,
"answers": [
{
"id": 1,
"content": "You can solve linear equations by using the substitution method.",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_comments": 2,
"number_of_votes": 1,
"comments": [
{
"id": 1,
"content": "That's a great answer!",
"likes": 1,
"dislikes": 0,
"number_of_clicks": 1,
"number_of_votes": 1,
}"
]
}
]
}
]
}
]
}
```
I want the client to be able to QUERY this in any specific way, maybe he wants GET objects with this format
```yml
{
"id": 1,
"name": "Math"
"tags": [
{
"id": 1,
"name": "Algebra",
}]
"topics": [
{
"id": 1,
"name": "Linear Equations",
}
]
}
```
But it could be totally different, and I can't predict the way the client is gonna query it and make a new route for every combination and possible schema for a complex object. The QUERY route would accept a request body, and then we would filter the response with the schema the client sends on request body.
- Would it be too complicated to implement this on FastAPI?
- Is there any hope that this will someday be implemented in FastAPI? XD
Thanks!
### Wanted Solution
...
### Wanted Code
```python
...
```
### Alternatives
...
### Operating System
macOS
### Operating System Details
_No response_
### FastAPI Version
A
### Python Version
3.11
### Additional Context
_No response_</div> | 0easy
|
Title: Integrate PostgreSQL 16 Support in Django Cookiecutter
Body: ## Description
I propose to add PostgreSQL 16 as a selectable option for the database setup within the Django Cookiecutter project template. This update would necessitate adjustments to Docker configuration files, database connectivity settings, and potentially updating version-specific dependencies to ensure compatibility. Incorporating PostgreSQL 16 will enable users to leverage the latest improvements in database functionality, performance enhancements, and security features offered by the most recent version of PostgreSQL.
## Rationale
The integration of PostgreSQL 16 into the Django Cookiecutter template is essential to keep the template up-to-date with current database technologies, providing users with the best possible tools for their projects. This addition will cater to developers looking to use the latest database features and optimizations in their new projects, enhancing the template's appeal and utility. Furthermore, staying current with database versions can improve application security and performance, aligning with best practices in software development and deployment. | 0easy
|
Title: Add IsYearStart primitive
Body: - This primitive returns the `is_year_start` on a datetime column | 0easy
|
Title: Support for frozen dataclasses
Body: **Is your feature request related to a problem? Please describe.**
Not really a problem per say. But many of our current models are frozen dataclasses. We try to keep them immutable by default. But unfortunately theres not a flag for it when generating through this tool.
We will probably still explore generating these just as not frozen to start, so not really a blocker. But it would be awesome if we could bring that in soon.
**Describe the solution you'd like**
It seems like it would be straightforward to utilize the already available `--enable-faux-immutability` flag.
**Describe alternatives you've considered**
Could add another flag of `--frozen-dataclasses` or similar.
**Additional context**
Nothin that I can think of. If I have time I will try to get a PR up if desired :) | 0easy
|
Title: DVC log_plot error is confusing, if the `template` attribute is set without a value.
Body: # Bug Report
## Description
DVC `log_plot` error is confusing, if the `template` attribute is set without a value.
### Reproduce
1. Create a DVC repository
2. Create a sample experiment that uses the `log_plot` function to log a few metrics (Note: the `template` option is empty)
```python
import pandas as pd
from dvclive import Live
from sklearn.datasets import load_iris
iris = load_iris()
datapoints = pd.DataFrame(data=iris.data, columns=iris.feature_names)
with Live() as live:
live.log_plot(
"sepal",
datapoints,
x="sepal length (cm)",
y="sepal width (cm)",
template="",
title="Sepal width vs Sepal length")
```
3. Run the experiment
4. Error will be something similar to the one below
```bash
vscode โ /workspaces/processor/.dvc/tmp (feature/strategy3) $ dvc plots diff -v
2024-07-14 09:57:01,538 DEBUG: v3.51.2 (pip), CPython 3.10.13 on Linux-6.6.31-linuxkit-x86_64-with-glibc2.31
2024-07-14 09:57:01,538 DEBUG: command: /home/vscode/.local/bin/dvc plots diff -v
2024-07-14 09:57:03,690 ERROR: unexpected error - [Errno 21] Is a directory: '/workspaces/processor/.dvc/tmp'
Traceback (most recent call last):
File "/home/vscode/.local/lib/python3.10/site-packages/dvc/cli/__init__.py", line 211, in main
ret = cmd.do_run()
File "/home/vscode/.local/lib/python3.10/site-packages/dvc/cli/command.py", line 27, in do_run
return self.run()
File "/home/vscode/.local/lib/python3.10/site-packages/dvc/commands/plots.py", line 107, in run
renderers_with_errors = match_defs_renderers(
File "/home/vscode/.local/lib/python3.10/site-packages/dvc/render/match.py", line 131, in match_defs_renderers
renderer = renderer_cls(plot_datapoints, renderer_id, **first_props)
File "/home/vscode/.local/lib/python3.10/site-packages/dvc_render/vega.py", line 87, in __init__
self.template = get_template(
File "/home/vscode/.local/lib/python3.10/site-packages/dvc_render/vega_templates.py", line 724, in get_template
with _open(template_path, encoding="utf-8") as f:
IsADirectoryError: [Errno 21] Is a directory: '/workspaces/processor/.dvc/tmp'
2024-07-14 09:57:03,752 DEBUG: link type reflink is not available ([Errno 13] Permission denied: '/workspaces/.VcJKvhyzGauYU48vSH9qxQ.tmp')
2024-07-14 09:57:03,752 DEBUG: Removing '/workspaces/.VcJKvhyzGauYU48vSH9qxQ.tmp'
2024-07-14 09:57:03,753 DEBUG: link type hardlink is not available ([Errno 95] no more link types left to try out)
2024-07-14 09:57:03,753 DEBUG: Removing '/workspaces/.VcJKvhyzGauYU48vSH9qxQ.tmp'
2024-07-14 09:57:03,753 DEBUG: link type symlink is not available ([Errno 13] Permission denied: '/workspaces/processor/.dvc/cache/files/md5/.WaF1GsOTeDBjuMNIcuPZpw.tmp' -> '/workspaces/.VcJKvhyzGauYU48vSH9qxQ.tmp')
2024-07-14 09:57:03,753 DEBUG: Removing '/workspaces/.VcJKvhyzGauYU48vSH9qxQ.tmp'
2024-07-14 09:57:03,754 DEBUG: Removing '/workspaces/processor/.dvc/cache/files/md5/.WaF1GsOTeDBjuMNIcuPZpw.tmp'
2024-07-14 09:57:03,761 DEBUG: Version info for developers:
DVC version: 3.51.2 (pip)
-------------------------
Platform: Python 3.10.13 on Linux-6.6.31-linuxkit-x86_64-with-glibc2.31
Subprojects:
dvc_data = 3.15.1
dvc_objects = 5.1.0
dvc_render = 1.0.2
dvc_task = 0.4.0
scmrepo = 3.3.6
Supports:
http (aiohttp = 3.9.5, aiohttp-retry = 2.8.3),
https (aiohttp = 3.9.5, aiohttp-retry = 2.8.3),
s3 (s3fs = 2024.6.1, boto3 = 1.34.131)
Config:
Global: /home/vscode/.config/dvc
System: /etc/xdg/dvc
Cache types:
Cache directory: fakeowner on /run/host_mark/Users
Caches: local
Remotes: s3
Workspace directory: fakeowner on /run/host_mark/Users
Repo: dvc, git
Repo.site_cache_dir: /var/tmp/dvc/repo/3e826ad9ce4b0c5c7907b07975f63692
```
### Expected
Error should point out that a `template` needs to be specified if the attribute is specified.
### Environment information
**Output of `dvc doctor`:**
```console
DVC version: 3.51.2 (pip)
-------------------------
Platform: Python 3.10.13 on Linux-6.6.31-linuxkit-x86_64-with-glibc2.31
Subprojects:
dvc_data = 3.15.1
dvc_objects = 5.1.0
dvc_render = 1.0.2
dvc_task = 0.4.0
scmrepo = 3.3.6
Supports:
http (aiohttp = 3.9.5, aiohttp-retry = 2.8.3),
https (aiohttp = 3.9.5, aiohttp-retry = 2.8.3),
s3 (s3fs = 2024.6.1, boto3 = 1.34.131)
Config:
Global: /home/vscode/.config/dvc
System: /etc/xdg/dvc
Cache types:
Cache directory: fakeowner on /run/host_mark/Users
Caches: local
Remotes: s3
Workspace directory: fakeowner on /run/host_mark/Users
Repo: dvc, git
Repo.site_cache_dir: /var/tmp/dvc/repo/3e826ad9ce4b0c5c7907b07975f63692
```
**Additional Information (if any):**
<!--
Please check https://github.com/iterative/dvc/wiki/Debugging-DVC on ways to gather more information regarding the issue.
If applicable, please also provide a `--verbose` output of the command, eg: `dvc add --verbose`.
If the issue is regarding the performance, please attach the profiling information and the benchmark comparisons.
-->
| 0easy
|
Title: error_correction_pipeline cannot correct any error!
Body: ### System Info
Pandasai 2.0.32
### ๐ Describe the bug
In **pandasai/pipelines/chat/code_execution.py
I find the following code
```python
def execute(self, input: Any, **kwargs):
...
while retry_count <= self.context.config.max_retries:
try:
result = self.execute_code(input, code_context)
...
except Exception as e:
...
code_to_run = self._retry_run_code(
code_to_run, self.context, self.logger, e
)
```
It appears that Pandasai intends to correct the error by updating code_to_run. However, in each iteration of the `while` loop, the code attempts to execute: `result = self.execute_code(input, code_context)`, where `input` consistently represents the initial code generated by the language model. Consequently, Pandasai never actually corrects its error.
Furthermore, in pandasai/pipelines/chat/error_correction_pipeline/error_correction_pipeline.py, the structure of the error_correction_pipeline is defined as follows:
```python
self.pipeline = Pipeline(
context=context,
logger=logger,
query_exec_tracker=query_exec_tracker,
steps=[
ErrorPromptGeneration(on_prompt_generation=on_prompt_generation),
CodeGenerator(),
CodeCleaning(),
],
)
```
The `CodeGenerator()` step lacks a callback parameter similar to `on_execution=on_code_generation`, which is utilized in `generate_chat_pipeline`. Consequently, Pandasai fails to record the newly corrected code. | 0easy
|
Title: Error with search trying tutorial
Body: I got an error while trying tutorial - after populating the database using fixtures any manage.py execution throws:
TypeError: 'RegexURLPattern' object is not iterable
This is somehow connected to search module - after deleting "search" page from cms all works.
Tried on ubuntu, both 16.10 and clean 16.04, docker image works right too.
| 0easy
|
Title: Monitoring: add Sentry
Body: Add Sentry configuration and test it. | 0easy
|
Title: Deprecate using arguments starting with `$`, `@`, `&` and `%` as literal values
Body: We are planning to support `$var` syntax in addition to `${var}` in the future (#4674). The plan was to do that already in RF 7.0, but it turned out that this would be too badly backwards incompatible. Probably the biggest problem would be that the Browser library accepts passwords in format `$secret` and resolves the variable value internally to avoid the value being logged automatically. After the planned change `$secret`, would be considered exactly the same as `${secret}` meaning that a confidential value could leak.
We still believe that supporting the `$var` syntax is a good idea, but we need to first deprecate using values starting with `$` or any other variable identifier. That means that using, for example,
```
Log Many $value @another
```
will cause a deprecation warning and the value needs to be escaped like
```
Log Many \$value \@another
```
| 0easy
|
Title: dfs `cutoff_time` should take a datetime string
Body: - As a user, I wish featuretools `dfs` would take a string as cutoff_time aswell as a datetime object
#### Code Example
```python
fm, features = ft.dfs(entityset=es,
target_dataframe_name='customers',
cutoff_time="2014-1-1 05:00",
instance_ids=[1],
cutoff_time_in_index=True)
```
as well as
```python
fm, features = ft.dfs(entityset=es,
target_dataframe_name='customers',
cutoff_time=dateutil.parser.parse("2014-1-1 05:00"),
instance_ids=[1],
cutoff_time_in_index=True)
```
| 0easy
|
Title: OSI Approved Licenses metric API
Body: The canonical definition is here: https://chaoss.community/?p=3962 | 0easy
|
Title: Customization of Precision Recall curves
Body: It would be best to add ability for the user to select which curves they want to display. E.g some people might not want to display the macro and micro averaged curves, display only a specific class' ROC curve.
This approach could then be extended to Precision Recall curves as well
EDIT: ROC Curves now has `curves` argument thanks to @doug-friedman | 0easy
|
Title: Python 3.12 compatibility
Body: Python 3.12 will be released in October and the first beta is now available. I tested it and noticed that our Python evaluation using the special variable syntax `$var` is broken. This seems to be due to a subtle change in Python's tokenizer (python/cpython#104802). Luckily fixing our code is easy and with the fix all our acceptance tests pass.
A much smaller issue is that our tests seem to have invalid escape sequences like `'\.'` that are nowadays reported more visibly than earlier. They need be changed to `'\\.'` or `r'\.'`. | 0easy
|
Title: masking function in multifit(mask) does not work as expected
Body: Dear all,
I would like to use the multifit function with a mask (see below). Unfortunately I get a ValueError message that makes no sense for me. Do I something wrong? Is there a workaround or a similar solution for my problem?
Thanks a lot!
```
s = hs.load("datasets/PS71-T_0013.dm3")
print(s.axes_manager)
<Axes manager, axes: (|512)>
Name | size | index | offset | scale | units
================ | ====== | ====== | ======= | ======= | ======
---------------- | ------ | ------ | ------- | ------- | ------
Energy loss | 512 | | 5.4e+02 | 0.8 | eV
```
`m = s.create_model()`
```
m.components
| Attribute Name | Component Name | Component Type
---- | ------------------- | ------------------- | -------------------
0 | PowerLaw | PowerLaw | PowerLaw
1 | Ni_L3 | Ni_L3 | EELSCLEdge
2 | Fe_L3 | Fe_L3 | EELSCLEdge
```
```
m.axes_manager._navigation_shape_in_array
()
```
```
mask = np.ones((1,512)).astype(bool)
mask[:,0:66] = False
mask[:,486:512] = False
mask.dtype
dtype('bool')
```
```
m.multifit(mask=mask, kind='smart')
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
Input In [273], in <module>
----> 1 m.multifit(mask=mask, kind='smart')
File D:\Anaconda\envs\hyperspy-eels\lib\site-packages\hyperspy\model.py:1569, in BaseModel.multifit(self, mask, fetch_only_fixed, autosave, autosave_every, show_progressbar, interactive_plot, iterpath, **kwargs)
1561 _logger.info(
1562 f"Autosaving every {autosave_every} pixels to {autosave_fn}.npz. "
1563 "When multifit finishes, this file will be deleted."
1564 )
1566 if mask is not None and (
1567 mask.shape != tuple(self.axes_manager._navigation_shape_in_array)
1568 ):
-> 1569 raise ValueError(
1570 "The mask must be a numpy array of boolean type with "
1571 f"shape: {self.axes_manager._navigation_shape_in_array}"
1572 )
1574 masked_elements = 0 if mask is None else mask.sum()
1575 maxval = self.axes_manager.navigation_size - masked_elements
ValueError: The mask must be a numpy array of boolean type with shape: ()
```
| 0easy
|
Title: Change behavior of `--test` and `--include` so that they are cumulative
Body: Currently when you use `--suite x --test y --include z`, a test is selected only if it is in suite `x`, has name `y` and contains a tag `z`. This is rarely if ever useful. It would typically be enough to just use `--test y` to select that test.
On the other hand, it would more useful to be able to use `--suite x --test y` so that all tests in suite `x` are selected in addition to test `y`. Similarly, `--include x --test y` should mean selecting all tests containing tag `x` in addition to test `y`.
This change is backwards incompatible, but because the current behavior is pretty strange it's unlikely there are lot of problems. Changing the behavior in Robot Framework 7.0 ought to thus be fine. The plan is to change the behavior with `--suite` [also](https://github.com/robotframework/robotframework/issues/4688) [otherwise](https://github.com/robotframework/robotframework/issues/4720) and making all these changes in the same release would be good.
Although these options should have a cumulative effect, we probably should interpret `--exclude` so that tests containing the specified tag aren't selected. That would preserve its behavior when used with other options, most importantly with `--include`.
---
**UPDATE:** We decided to keep the old `--suite` behavior. See the comment below for more details. Notice also that `--suite` not affecting which files are parsed was implemented already in RF 6.1 (#4688).
---
**UPDATE:** This change will be reverted in RF 7.0.1 due to problems it caused with Pabot and `--rerunfailed`. See comments and other notes below for more information. | 0easy
|
Title: Setting Swagger fields in `info`
Body: **Describe the bug**
Swagger provides the option to define many more fields under `info` such as:
- `termsOfService`
- `license`
- `license.name`
- `license.url`
- `contact`
- `contact.email`
There is no documented way to set these fields in the generated Swagger file.
**Expected behavior**
There should be a way to configure these fields.
**Python Information (please complete the following information):**
- Python v3.10.1
- Spectree v0.6.8
- Falcon v3.0.1
| 0easy
|
Title: Detect Duplicate Page URL Segments
Body: When creating multiple pages, either via `@rio.page`, `rio.ComponentPage` or `rio.Redirect`, it is possible for multiple pages to share a common URL segment. This is nonsensical, because it's not clear which one should be displayed when navigating to that URL.
This also causes secondary problems e.g. with the default navigation which doesn't expect duplicate URLs and then displays wrong links.
It would be nice to have a check for this, and either raise an exception. | 0easy
|
Title: Parsing error for topic9_part2_facebook_prophet.ipynb
Body: This notebook isn't rendering in nbviewer nor on Github nor after uploading it to binder..
https://nbviewer.jupyter.org/github/Yorko/mlcourse_open/blob/master/jupyter_english/topic09_time_series/topic9_part2_facebook_prophet.ipynb
there's an extra i at the beginning of the notebook(had checked the raw file)
Even removing that, the nbs isn't opening.. | 0easy
|
Title: Remove support for using HttpAuthMiddleware without http_auth_domain
Body: Deprecated in 2.5.1. | 0easy
|
Title: Add Warning in Docs that only timezone aware data is supported
Body: sqlalchemy.exc.DBAPIError: (sqlalchemy.dialects.postgresql.asyncpg.Error) <class 'asyncpg.exceptions.DataError'>: invalid input for query argument $3: datetime.datetime(2024, 5, 22, 8, 11, 40... (can't subtract offset-naive and offset-aware datetimes)
[SQL: UPDATE platform SET is_active=$1::BOOLEAN, is_live=$2::BOOLEAN, updated_at=$3::TIMESTAMP WITHOUT TIME ZONE WHERE platform.id = $4::INTEGER]
[parameters: (True, True, datetime.datetime(2024, 5, 22, 8, 11, 40, 307132, tzinfo=datetime.timezone.utc), 2)]
showing this error when i try to update my table.
The issue i think is in my tables the field updated_at didn't require timezone but the update code in fastcrud adding timezone in it so how to avoid that? | 0easy
|
Title: InputObjectType fields from models
Body: I've used a `MongoengineObjectType` class to define the schema from model.
There's a way to use a reference model to define a `InputObjectType` or `AbstractType` class? Something like that:
```
class UserFields(graphene.AbstractType):
class Meta:
model = model.User
class User(graphene.ObjectType, UserFields):
pass
class UserInput(graphene.InputObjectType, UserFields):
pass
``` | 0easy
|
Title: [BUG] TinyTimeMixer should not create validation set if validation_split is None
Body: **Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
`TinyTimeMixerForecaster` always creates a validation set even if `validation_split is None`, in which case the `temporal_train_test_split` defaults the `validation_split` to `0.25`
Following fixes should be made:
- create `y_test` only if `validation_split is not None`
- rename data split: `test` to `eval`
**To Reproduce**
<!--
Add a Minimal, Complete, and Verifiable example (for more details, see e.g. https://stackoverflow.com/help/mcve
If the code is too long, feel free to put it in a public gist and link it in the issue: https://gist.github.com
-->
```python
from sktime.forecasting.ttm import TinyTimeMixerForecaster
from sktime.datasets import load_airline
y = load_airline()
forecaster = TinyTimeMixerForecaster(
config={"context_length": 10},
training_args={
"output_dir": "test_output",
"report_to": "none",
},
validation_split=None, # defaults validation_split to 0.25
)
forecaster.fit(y, fh=[1, 2, 3])
y_pred = forecaster.predict()
```
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
when `validation_split is None`, `test` should not be created and not passed to the `Trainer`
**Versions**
<details>
<!--
Please run the following code snippet and paste the output here:
from sktime import show_versions; show_versions()
-->
```
System:
python: 3.12.5 | packaged by Anaconda, Inc. | (main, Sep 12 2024, 18:27:27) [GCC 11.2.0]
executable: /home/geetu/miniconda3/envs/ai/bin/python
machine: Linux-6.8.0-51-generic-x86_64-with-glibc2.39
Python dependencies:
pip: 24.2
sktime: 0.34.0
sklearn: 1.5.2
skbase: 0.8.3
numpy: 1.26.4
scipy: 1.14.1
pandas: 2.2.3
matplotlib: 3.9.2
joblib: 1.4.2
numba: None
statsmodels: None
pmdarima: None
statsforecast: None
tsfresh: None
tslearn: None
torch: 2.4.1
tensorflow: 2.18.0
```
</details>
<!-- Thanks for contributing! -->
| 0easy
|
Title: Is changing `Power` from an NpElemOperator to an NpPairOperator a good idea?
Body: In order to achieve the expression like **"Power($close, Rank($close, 14))"** which is fairly common in fomulla alpha.
I guess we could use following code:
`class Power(NpPairOperator):
def __init__(self, feature_left, feature_right):
super(PairPower, self).__init__(feature_left, feature_right, "power")`
Finally, we could add a __rpow__ magic method to **Expression** class | 0easy
|
Title: Linreg does not return Slope, Forecast, Angle, or R coefficient
Body: Which version are you running? 0.3.14b0
Do you have TA Lib also installed in your environment? No
The LINREG indicator returns the linreg indicator value but does not return other useful indicator values mentioned in the documentation - e.g. Slope, Forecast, Angle, or R coefficient which are calculated in the indicator code.
Sample code to reproduce:
```python
import pandas as pd
import pandas_ta as ta
import yfinance as yf
print(f"PANDAS TA version = {ta.version}")
# Get symbol OHLC data
data = yf.download("CL=F")
# pandas_ta linreg doesn't appear to be returning all the values yet...
# this works
data['LINREG'] = data.ta.linreg(length=20, slope=True, degrees=True, r=True, tsf = True )
# this (below) doesn't work as dictionary key values are missing
# linreg = data.ta.linreg(length=20, slope=True, degrees=True, r=True, tsf = True )
# data['LINREG'] = linreg['LR_20']
# data['LINREG_SLOPE'] = linreg['LRm_20']
# data['LINREG_FORECAST'] = linreg['LR_20']
# data['LINREG_ANGLE'] = linreg['LRa_20']
# data['LINREG_R'] = linreg['LRr_20']
```
Expected: indicator values described in the documentation (and calculated in the indicator code) should be available | 0easy
|
Title: bson.errors.InvalidDocument: cannot encode object: frozenset()/set()
Body: Exception while saving an object containing either a frozenset or a set:
```py
from odmantic import Model, AIOEngine
from odmantic.bson import ObjectId
class UserModel(Model):
username: str
favorites_article_ids: FrozenSet[ObjectId] = frozenset()
engine = AIOEngine()
await engine.save(UserModel(username="jean"))
``` | 0easy
|
Title: Remove office hours announcements
Body: ### Summary
Remove this:
<img width="817" alt="Image" src="https://github.com/user-attachments/assets/ce9c4e1b-3963-40e1-bd1d-70c77f05a443" />
### Where to change
```diff
diff --git a/README.md b/README.md
index 6d2af8a80e..8d42c95990 100644
--- a/README.md
+++ b/README.md
@@ -1,13 +1,5 @@
# MLflow: A Machine Learning Lifecycle Platform
-### ๐ข Announcements ๐ข
-
-### [Join us at the upcoming office hours (Feb 5, 2025 5:00 PM PST / 8:00 PM EST)](https://lu.ma/gfcdbuz4)
-
-To view all upcoming MLflow community events, check out [lu.ma/mlflow](https://lu.ma/mlflow)
-
----
-
[](https://mlflow.org/docs/latest/index.html)
[](https://github.com/mlflow/mlflow/blob/master/LICENSE.txt)
[](https://pepy.tech/project/mlflow)
```
### Notes
- Make sure to open a PR from a **non-master** branch.
- Sign off the commit using the `-s` flag when making a commit:
```sh
git commit -s -m "..."
# ^^ make sure to use this
```
- Include `#{issue_number}` (e.g. `#123`) in the PR description when opening a PR.
| 0easy
|
Title: Documentation may have code error!
Body: <!--
**IMPORTANT**:
- Use the [Gensim mailing list](https://groups.google.com/forum/#!forum/gensim) to ask general or usage questions. Github issues are only for bug reports.
- Check [Recipes&FAQ](https://github.com/RaRe-Technologies/gensim/wiki/Recipes-&-FAQ) first for common answers.
Github bug reports that do not include relevant information and context will be closed without an answer. Thanks!
-->
#### Problem description
[Similarity Queries](https://radimrehurek.com/gensim/auto_examples/core/run_similarity_queries.html#sphx-glr-auto-examples-core-run-similarity-queries-py)--Part--`Performing queries`
I think there is a code error. Document id error. Please see the code below.
### code
```python
sims = sorted(enumerate(sims), key=lambda item: -item[1])
for i, s in enumerate(sims):
print(s, documents[i]) # s[0] is the document id, the 'i' is not.
```
corrected code should be
```python
sims = sorted(enumerate(sims), key=lambda item: -item[1])
for i, s in enumerate(sims):
print(s, documents[s[0]])
```
| 0easy
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.