-
Notifications
You must be signed in to change notification settings - Fork 86
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unsuccessful Build on A10-7850K, please help! #65
Comments
You are not doing anything wrong. There is an issue with SYCL and the SIMD instructions (i.e. mavx and msee flags). I am running tests right now to see if a simple fix is possible which would allow to keep the SIMD instructions for the CPU. Also I noticed that you are using clang. You should switch to gcc-4.8 (and g++4.8) at least. |
Ok this seems to compile and pass the tests just fine. |
Thank you. That helped with compiling! Compiling now finishes successfully. However, I am not quite sure how to proceed from here to get a working build within a virtual environment. I tried to create a wheel like so:
Then, I tried:
Now when I try to just import Tensorflow in Jupyter Notebook, I am getting the following errors:
Am I too ambitious trying to make a wheel and then do a pip install into my virtual environment? I see that in the tutorial your colleague is actually: ` $ cd _python_build $ ln -s ../bazel-bin/tensorflow/tools/pip_package/build_pip_package.runfiles/org_tensorflow/* . $ ln -s ../tensorflow/tools/pip_package/* . |
Ok so this is just some incomplete information. Also I wasn't able to create a whl like you tried either. Not sure why. |
Hi Zakor, I thought it would be wise for me to also run the tests as recommended. I am noticing 2 things: 1.) There is a timeout option set on the tutorial like so: It's rather high. Are these limits high because the test computations are rather expensive? 2.) A lot of these tests time out on my setup. Is this an indication that the integrated GPU is not quite powerful enough to perform these tests in time? At the writing of these tests, the tests have been running for 2 days now. I expect the tests to finish within one more day. Is this an indication of deeper problems? E.g.: The GPU is not actually used etc. Update: After compiling the build without errors, I proceeded to compare some performance: Tensorflow from pip: 1 epoch = ~ 1,700 sec (CPU utilization ~ 350%) This is about 10x worse than what I would get with just optimized tensorflow compilation. I am not currently gaining the hoped for performance increases. How can I test if the GPU is used at all? |
Hi, |
Thank you for guiding me along: I have reinstalled tensorflow-opencl because I specified incorrect paths to the computcpp, g++ and g compiler last time. I have also rerun my CIFAR-10 dataset to benchmark tensorflow-opencl with the integrated GPU on my AMD A10-7850 setup. This time I also looked at GPU usage. I was able to verify that the GPU is used at 100% capacity. The CPU usage is still at 150%. The time to process 1 epoch is still about 5800 seconds . So this is still about 5X worse than just using optimized compilation flags. These are my general observations and thoughts after having tried out tensorflow-opencl: 1.) The fact that on my 4-core setup, the CPU usage is only 150% percent indicates that somehow multithreading is not fully efficient in the opencl setup because without opencl all 4 cores get used. How can I further increase computation times without NVIDIA GPUs? I still have a few months before I commit. One of the things I am waiting for is the release of the VEGA GPUs. In the absence of any softwareframe work ontop to exploit the GPU potential, it will be hard to go with the VEGA cards though. For one, the talked about rocM and MIopen initiatives have not been released. In general what are your thoughts? (Oh and by the way, if you need help with benchmarking on an APU system, now that things are working for me, I would be happy to help out!) |
Well assuming you don't want to spend more money on GPU, the only possibility I see is contributing to this repository ;) There are other optimizations to do to avoid copies. (ok that's very nice of you ^^) |
Well, I will spend more money on a GPU in about 2-5 months. The question then will be what GPU to get (NVIDIA vs AMD). At this point, the answer is tilting toward NVIDIA. But I really want to give AMD a good chance first. With MIOpen and rocM, and this, the most robust path on AMD GPUs seems less defined. This probably goes outside the scope of this discussion...in which case, could you point me to lists of optimizations that still need to be done? |
Yes this is definitely getting out of scope. Please open an issue on https://github.com/lukeiwanski/tensorflow about the optimizations that can be done to avoid copies. |
This is a follow-up on a previous message. I am encountering build errors, and don't seem to be able to find the source of it.
I have followed the following steps that I believe your colleague posted here:
https://www.codeplay.com/portal/03-30-17-setting-up-tensorflow-with-opencl-using-sycl
I deviated these instructions in the following way:
I did not update execute the following steps:
I was not sure why it is important to go to a that particular kernal so I did not upgrade the kernel. This is the version of Ubuntu I am using:
Distributor ID: Ubuntu
Description: Ubuntu 14.04.5 LTS
Release: 14.04
Codename: trusty
I am using the following kernel as part of t his standard Ubuntu 14.04.5 built:
3.13.0-116-generic
I used Python 3.5 inside a conda environment instead of Python 2.7
clinfo gives the following info:
/usr/local/computecpp/bin/computecpp_info gives the following output
I note here that somehow the CPU was not detected which is different from the tutorial mentioned above.
After configuring with default options, I run the following command:
I am encountering the following error:
Did skipping some of the steps outlined above really lead to these errors? What did I do wrong?
The text was updated successfully, but these errors were encountered: