superset
superset copied to clipboard
docker-compose up and can't open the UI
A clear and concise description of what the bug is.
How to reproduce the bug
According this doc
- git clone
- cd superset
- touch ./docker/requirements-local.txt and echo
clickhouse-connect
> ./docker/requirements-local.txt - docker-compose up
Expected results
Open the superset in browser.
Actual results
Can not open superset in browser like this:
Screenshots
![image](https://user-images.githubusercontent.com/7600925/223592578-18736dc4-91a4-4024-b3f1-51b23fcaf48d.png)
Environment
(please complete the following information):
- browser type and version: Chrome
- superset version:
the newest code
- python version:
python3.9
- node.js version:
v14.19.3
Checklist
Make sure to follow these steps before submitting your issue - thank you!
- [x] I have checked the superset logs for python stacktraces and included it here as text if there are any.
- [x] I have reproduced the issue with at least the latest released version of superset.
- [x] I have checked the issue tracker for the same issue and I haven't found one similar.
Additional context
werkzeug.exceptions.NotFound: 404 Not Found: The requested URL was not found on the server. If you entered the URL manually please check your spelling and try again.
superset_app | 172.19.0.1 - - [08/Mar/2023 01:11:59] "GET /static/assets/images/favicon.png HTTP/1.1" 404 -
superset_app | 2023-03-08 01:11:59,855:INFO:werkzeug:172.19.0.1 - - [08/Mar/2023 01:11:59] "GET /static/assets/images/favicon.png HTTP/1.1" 404 -
superset_worker_beat | [2023-03-08 01:12:00,094: INFO/MainProcess] Scheduler: Sending due task reports.scheduler (reports.scheduler)
superset_worker | [2023-03-08 01:12:00,174: INFO/MainProcess] Task reports.scheduler[3626bbb8-062a-49e7-8f2a-033ed05decf9] received
superset_worker | [2023-03-08 01:12:00,632: INFO/ForkPoolWorker-3] Task reports.scheduler[3626bbb8-062a-49e7-8f2a-033ed05decf9] succeeded in 0.4056195000000571s: None
superset_app | 2023-03-08 01:12:00,899:WARNING:superset.views.base:HTTPException
superset_app | Traceback (most recent call last):
superset_app | File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1517, in full_dispatch_request
superset_app | rv = self.dispatch_request()
superset_app | File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 1503, in dispatch_request
superset_app | return self.ensure_sync(self.view_functions[rule.endpoint])(**req.view_args)
superset_app | File "/usr/local/lib/python3.8/site-packages/flask/app.py", line 509, in <lambda>
superset_app | view_func=lambda **kw: self_ref().send_static_file(**kw), # type: ignore # noqa: B950
superset_app | File "/usr/local/lib/python3.8/site-packages/flask/scaffold.py", line 322, in send_static_file
superset_app | return send_from_directory(
superset_app | File "/usr/local/lib/python3.8/site-packages/flask/helpers.py", line 676, in send_from_directory
superset_app | return werkzeug.utils.send_from_directory( # type: ignore
superset_app | File "/usr/local/lib/python3.8/site-packages/werkzeug/utils.py", line 586, in send_from_directory
superset_app | raise NotFound()
superset_app | werkzeug.exceptions.NotFound: 404 Not Found: The requested URL was not found on the server. If you entered the URL manually please check your spelling and try again.
The same problem.
The same problem.
i'm having this same issue
I'm having te same problem.
The same issue. Can't start in docker-compose
same here... any hits on what to do?
I also faced with such problem, but deleting the old node_modules
folders helped in my case.
You may use one of this commands:
-
git clean -xdf
-
git reset --hard
<- 🛑 DANGER, COMMIT LOCAL CHANGES BEFORE!
Looks like the problem is caused by permissions to node_modules
folder. It is owner by root, not by user used to run superset. Changing the ownership fixed the issue for me. Actually, I changed it for all the files:
chown -R myuser:mygroup /path_to_superset
node_modules
@bukem thank you your solution works, also for lazy people like me :) they can also try:
chmod -v -R 777 node_modules/
Having the same issue, none of the proposed solutions work for me... :(
The original report is using node 14. Can you try with node 16? We're actually going to be on node 18 imminently, and moving to 20, so 14 is probably old enough to be quite problematic.