We are porting Azure OS https://github.com/azure-rtos/threadx/tree/v6.1.11_rel
to TMS570 hardware https://github.com/azure-rtos/threadx/tree/v6.1.11_rel/ports/cortex_r5.
We have requirement of floating point calculations on application side. Can anyone tell me how do we can enable floating point support.
Make sure floating point is enabled in the compiler and assembler options when you build ThreadX and your application. For each thread using the VFP, call tx_thread_vfp_enable() before using the VFP.
Related
I am new to QEMU, but to support a system modeling project, I try to figure out how to emulate a new ARM microcontroller(M33 based) SOC which is not on the supported system list. I have checked on the QEMU documentation: https://qemu-project.gitlab.io/qemu/about/index.html but was not able to find directions to construct and add a new system model. I would appreciate very much if someone can point me to the right documentation, or suggest steps I can follow to generate the new model for emulation.
I would highly suggest using Renode rather than Qemu as it's better suited for microcontrollers and it's well documented and actively maintained on GitHub.
There is no documentation for how to write new machine models for QEMU. The best advice is to look at the source code for an existing machine type that seems similar to the one you want to add. Avoid looking at machine types that have been added to QEMU a long time ago -- they are often written in an older style or using older APIs which are no longer recommended for new code.
As a very rough rule of thumb, you should assume that a new machine model is about as much effort as porting an operating system to that new hardware. (Roughly, a device model is about as much code as a device driver.) Depending on how much functionality you need, you may be able to leave some or many devices in your SoC unimplemented -- the absolute minimum is usually something like a UART and a timer. You'll need documentation of the SoC (technical reference manual or similar, with detailed descriptions of the devices down to the register level).
(Out of curiosity, what is the M33-based SoC?)
What would be my best bet for trying to roll my own libraries when I want to launch Minecraft on an ARM device? Standard Mojang repositories only have x86 versions of LWJGL in particular, and trying to overwrite the copies in .minecraft/libraries with armhf version triggered a corrupt file error and a redownload cycle with the x86 version.
I think you are wanting to run Minecraft on mobile? Correct me if this is incorrect, Because LWJGL is already compatible with ARM, it doesn't even care which CPU it is running on (e.g AMD CPUs are ARM) because the two main APIs it is using (Java and OpenGL) both don't either - It even supports all OSs because of this too, it's just preferred that you use 64-bit libraries on 64-bit systems for speed purposes, you can always just use 32-bit even.
I would also like to clear up that Mojang did not make LWJGL so searching their repositories isn't the place to look. Plus if it were possible to simply change the natives they would have done that instead of make it again..
I am trying to build a fairly simple sqlite database based mobile app using mvvmcross and Portable class libraries. The database I have running is fairly large so querying it takes enough time where I don't want the UI to get blocked while running queries.
The way I currently have it set up is in a few classes based on the mvvmcross n+1 tutorials n=10 tutorial. I have two services that manage the look-ups for the two entities.
How can I perform these database calls on a separate thread and have the view be updated when completed. I assume that this capability exists within mvvmcross I just haven't been able to track down the documentation or any tutorials on it specifically.
Any help pointing me to the right direction would be much appreciated.
Thanks,
JH
Portable class libraries do give you access to the ThreadPool - so you could use that to marshall the work onto a background thread? http://msdn.microsoft.com/en-us/library/system.threading.threadpool.queueuserworkitem.aspx
Alternatively, if you have a setup/configuration which allows the TPL to be used in your library, then you could use the TaskFactory or async and await to provide some background threading - this is probably the better long-term route, but getting this setup and working initially may take longer as Xamarin's async support is very new and it's PCL support is changing.
Does anyone have any experience plugging into an Amazon gpu cluster in a web app? Is it even possible? I want to make an online interface to a simulator I have written in cuda, but everything I have seen so far involves ssh and the command line.
Yes, this is very possible!
You may want to consider binding your CUDA C code (I'm just assuming C) to your preferred scripting language for better efficiency if you intend to build the front end logic not in a lower level programming language but absolutely this is possible.
Keep in mind, many web developers build their apps strictly from the command line, you simply need to find an effective and efficient way of calling the CUDA code from your Web app and then simply return the results to the client who made the request.
I came across this problem,I have got an algorithm that I need to implement on websites.
The server side scripts may differ and it can be PHP,ASP.NET etc.
All I need to do to is to deliver the binary(I need it as I dont want this security algorithm to be open and viewable) that can comply by every type.As per me solution can be (Please correct me if I am wrong):-
Implementing binary according to OS.
Implementing Algo as per every type of script.(Tough and less portable)
Please suggest if there is other way round or please close this question and redirect me to any earlier question asked for this situation.I am new for this.
Thanks
I would suggest you to use Java Server-side technologies to implement your algorithm. You can write your algorithm as a java class, which can be called from a Servlet or a JSP or even any other technologies over http protocol. The main reason why I suggest using java is:
1) it is platform independent, so your 1st point:
Implementing binary according to OS.
You dont have to worry which OS the client would use.and it can be ported to other OS very easily.
and
2) it will be very secure, once you compile, a class file will be generated, which can be delivered. it cant be opened and viewable.