McHale
Regular
Hi @Space Cadet been a busy week so slow response and Go Lions here tooGo the Lions
SC
Hi @Space Cadet been a busy week so slow response and Go Lions here tooGo the Lions
SC
Not boasting just proving there are a few of us lucky ones!
View attachment 70069
I also hope that she is healthy again.
I see too much speculation and guesswork in your statement!
She mentions that she has a WhatsApp group that has a few members.
She writes that she knows Thomas Hulsing - possibly via LinkedIn or maybe even privately.
In general, I still don't see any proof that Thomas Hulsing is supposed to be in the WhatsApp group.
You can continue to speculate - maybe someone will dare to contact Thomas Hulsing.
Thank you!
Today we are celebrating our reunification here in Germany, no more wall, no wall, no wall, break down the wall - break down the wall BREAK DOWN THE WALL
DIE MAUER MUSS WEG!
I find two other things from you DU extremely interesting.
One is Shiraz gin and the other, even more fascinating, is Shiraz sparkling wine.
Break down the wall
___
Sorry!
...not to forget Vegemite
View attachment 70311
This was just posted on a Brainchip Facebook page, not to sure if it's been posted here before.
Sorry but there was no link on the post from the Facebook pageCan you also post the link please.
Sorry but there was no link on the post from the Facebook page
Actually both of them have nothing to celebrate… especially the East Germans. When they opened the border, the west Germans made a lot of money because of them… they bought everything overpriced from cars which was already ready to be demolished but they baught it because it was western car… to electric devices… even their savings from the bank was overnight almost worthless…and now they can’t even like each other…. Yeah… happy reunionYou are either from East Germany which used to be behind the wall, only those Germans really celebrate that day, or you opened the second bottle of wine already
Pretty sure this was from their annual report.View attachment 70312
Here's a copy of it without the volume bar in it . Once again this is a copy from someone's post on a Facebook group
I have already contacted him and asked. He read the discussion about the WhatsApp group and then replied to me.If you do not believe what you were told, contact Hülsing yourself! Where is the problem?
He is an investor just like us!
Pretty sure this was from their annual report.
If you go to the Tata website you will be able to read all about it. Beware, it is substantial.
Sorry but there was no link on the post from the Facebook page
I should really go to bed soon, but I just keep on finding Brainchip-related nuggets online today - believe it or not, even a job ad from Ukraine !
Junior Machine Learning Engineer at Data Science UA
**About us:**<br>Data Science UA is a service company with strong data science and AI expertise. Our journey began in 2016 with the organization of Kyiv’s first technical Data Science UA Conference,...djinni.co
Junior Machine Learning Engineer
Data Science UA
Diana Marchenko, IT Recruiter
About us:
We are Data Science UA, and we are a fast-growing IT service company. We are proud of developing the Data Science community in Ukraine for more than 7 years. Data Science UA unites all researchers, engineers, and developers around Data Science and related areas. We conduct events on machine learning, computer vision, intelligence, information science, and the use of artificial intelligence for business in various fields.
About role:
Data Science UA is looking for a Junior Machine Learning Engineer to become a helping hand for our internal Data Science team. Do you want to work on some real projects for our Clients and/or perform R&D of novel AI algorithms and platforms (in particular, on the neuromorphic platform Brainchip Akida)? Then apply and join our team! You will report directly to our Head of AI Consulting, PhD, which is a great opportunity for you and your further growth.
Requirements:
0,5-1 year of experience as ML Engineer/Data Scientist or related (alternatively, participation in open-source ML projects or Kaggle competitions);
Minimal expertise in CV or NLP-based projects;
Good knowledge of math, CS and AI fundamentals;
Proficiency in Python;
Student/graduate in the field of exact sciences (computer science, mathematics, cybernetics, physics, etc.);
Intermediate English.
Nice to have:
✔Have your own pet projects;
✔Completed different Data Science related courses.
Responsibilities:
Participation in AI consulting projects for Clients from Ukraine and abroad;
Work on complex R&D projects;
Write and publish scientific papers (and possibly your diploma in University);
Preparation of technical content (presentations, analytical articles).
We offer:
Opportunity to grow and participate in large projects of our AI R&D centers;
Possibility of remote work;
Development of professional skills and support in acquiring new knowledge (by attending conferences on Data Science, exchange of experience, etc.);
Friendly team;
Interesting tasks.
About Data Science UA
We understand that you need more than just a search engine to find IT and technical professionals or to progress your own career, that is why we established Data Scientist UA.
One place connecting business and developers.
Company website:
https://data-science-ua.com/
DOU company page:
https://jobs.dou.ua/companies/data-science-ua/
Job posted on 29 November 2023
50 views 15 applications
Pfftt..Just over ten months ago, I spotted the above job ad by Data Science UA, a company from Ukraine:
“(…) Data Science UA is looking for a Junior Machine Learning Engineer to become a helping hand for our internal Data Science team. Do you want to work on some real projects for our Clients and/or perform R&D of novel AI algorithms and platforms (in particular, on the neuromorphic platform Brainchip Akida)? (…)”
I just noticed that earlier today their Head of AI Consulting, who has a PhD in polymer chemistry and describes his job as “specializing in integrating AI solutions into chemical and pharmaceutical R&D” posted very favourably about neuromorphic chips in general and Akida in particular, which they had assessed in their latest R&D project, benchmarking it against NVIDIA Jetson hardware. A paper on this R&D project is currently being worked on.
View attachment 70330
View attachment 70332
View attachment 70333
Nice to see we make the list and for good reason too imo. The authors final thoughts appear to concur with what's needed and what BRN (& obviously other smaller providers) are offering against the behemoths and their power sucking, water evaporating solutions.
We know we play in the in-cabin space but interesting it was highlighted as a specialisation
A few enterprise takeaways from the AI hardware and edge AI summit 2024 - DataScienceCentral.com
Image by Gerd Altmann from Pixabay A few enterprise takeaways from the AI hardware and edge AI summit 2024 Enterprises haven’t seemed as enthusiastic about generative AI and large language models (LLMs) lately as they have been in previous years. The Kisaco Research event I attended in September...www.datasciencecentral.com
Alan Morrison is an independent consultant and freelance writer on data tech and enterprise transformation. He is a contributor to Data Science Central with over 35 years of experience as an analyst, researcher, writer, editor and technology trends forecaster, including 20 years in emerging tech R&D at PwC.
A few enterprise takeaways from the AI hardware and edge AI summit 2024
Alan Morrison- October 1, 2024 at 10:09 am
Image by Gerd Altmann from Pixabay
A few enterprise takeaways from the AI hardware and edge AI summit 2024
Enterprises haven’t seemed as enthusiastic about generative AI and large language models (LLMs) lately as they have been in previous years. The Kisaco Research event I attended in September provided some reasons why.
Current gen AI processing far too centralized to be efficient or cost effective
If there’s a single takeaway that I could point to, it’s how overloaded data centers are and how limited edge infrastructure has been when it comes to effectively reducing that data center load for gen AI applications. Ankur Gupta, Senior Vice President and General Manager at Siemens Electronic Data Automation noted during his talk that “the opportunity for low power needs to be met at the edge.”
Gen AI-oriented data centers must handle an inordinate amount of heat per GPU. Gupta asserted that half a liter of water evaporates with every ChatGPT prompt.
The newest, largest GPUs run even hotter. Tobias Mann in The Register in March 2024 wrote that “Nvidia says the [Blackwell] chip can output 1,200W of thermal energy when pumping out the full 20 petaFLOPS of FP4.” Even so, Charlotte Trueman writing in August 2024 in Data Center Dynamics and citing Nvidia CFO Colette Kress, wrote that “Nvidia was expecting to ship ‘several billion dollars in Blackwell revenue during Q4 2024.’”
Edge infrastructure innovation is key, with significant spending planned. IDC recently estimated that global spending on edge computing will reach $228 billion in 2024, a 14% increase from 2023. IDC forecasts spending to rise to $378 billion by 2028, a 13 percent CAGR from 2024 levels.
The research firm “expects all 19 enterprise industries profiled in the spending guide to see five-year double-digit compound annual growth rates (CAGRs) over the forecast period,” with the banking sector spending the most, according to CDO Trends.
To justify this level of investment, Lip-Bu Tan, chairman of Walden International, forecast edge AI revenue potential of $140 billion annually by 2033.
Aren’t smaller language models (SLMs) better for most purposes?
Donald Thompson, Distinguished Engineer @ Microsoft / LinkedIn, compared and contrasted LLMs with SLMs, saying he favors SLMs. It’s not often, he says, that users really need state-of-the-art LLMs. SLMs can allow faster inference, more efficiency and customizability.
Moreover, a solid micro-prompting approach can harness the power of functional, logically divided tasks, in the process improving accuracy. Thompson shared an example user dialogue agentic workflow that enables a form of knowledge graph creation. A dialectic that’s part of this user dialogue flow includes thesis, antithesis and synthesis, eliciting a broader, more informed viewpoint.
Pragmatic enterprise AI starts with better data and organizational change management
Manish Patel, Founding Partner at Nava Ventures, moderated a panel session on
“Emerging Architectures for Applications Using LLMs – The Transition to LLM Agents.” Panelists included Daniel Wu of the Stanford University AI Professional Program, Arun Nandi, Senior Director and Head of Data & Analytics, at Unilever, and Neeraj Kumar, Chief Data Scientist at Pacific Northwest National Laboratory.
The prospect of agentic AI is placing much more focus on the need for governance, risk assessment and improved data quality.
In order for those improvements to be realized, AI adoption must wait for organizational change. Wu pointed out that inside enterprises, “Change management is the single point of failure.” Even successful change efforts take years.
Moreover, expectations about AI are often unrealistic, with executives who don’t have the patience to wait for return on investment.
Kumar underscored the cross-functional nature of AI deployments and ownership considerations that arise as a result.
Nandi figured that “70 percent of the effort (in enterprise AI initiatives)” are change management related, and that such initiatives imply a need for much more extensive collaboration, given AI’s cross-functional nature, with the right people in the right roles in the loop.
Effective edge AI requires a different, linear modeling approach
Stephen Brightfield, CMO of neuromorphic IP provider Brainchip, presented on “Combining Efficient Models with Efficient Architectures.” Brainchip specializes in on-chip, in-cabin processing technology for smart car applications.
Brightfield asserted that “most edge hardware is stalled because it’s designed with a data center mentality.” Some of the observations he made underscored the learnings of an edge-constrained environment:
Rather than stick with a transformer-based neural net of the kind used in LLMs, Brainchip advocates a state space, state evolving, event-based model that promises more efficiency and lower latency.
- Assume fixed power limits.
- Lots of parameters implies a lot of data to move.
- Most data isn’t relevant.
- Sparse data implies more efficiency.
- Don’t recompute what hasn’t changed.
A final thought
Much media coverage is focused on LLM behemoths and the data center-related activities of hyperscalers. But what I found much more compelling were the innovations of smaller providers who were trying to boost the performance and utility of edge AI. After all, inferencing is consuming 80 percent of the energy AI demands, and the potential clearly exists to improve efficiencies through better and more pervasive edge processing.
Nice to see we make the list and for good reason too imo. The authors final thoughts appear to concur with what's needed and what BRN (& obviously other smaller providers) are offering against the behemoths and their power sucking, water evaporating solutions.
We know we play in the in-cabin space but interesting it was highlighted as a specialisation
A few enterprise takeaways from the AI hardware and edge AI summit 2024 - DataScienceCentral.com
Image by Gerd Altmann from Pixabay A few enterprise takeaways from the AI hardware and edge AI summit 2024 Enterprises haven’t seemed as enthusiastic about generative AI and large language models (LLMs) lately as they have been in previous years. The Kisaco Research event I attended in September...www.datasciencecentral.com
Alan Morrison is an independent consultant and freelance writer on data tech and enterprise transformation. He is a contributor to Data Science Central with over 35 years of experience as an analyst, researcher, writer, editor and technology trends forecaster, including 20 years in emerging tech R&D at PwC.
A few enterprise takeaways from the AI hardware and edge AI summit 2024
Alan Morrison- October 1, 2024 at 10:09 am
Image by Gerd Altmann from Pixabay
A few enterprise takeaways from the AI hardware and edge AI summit 2024
Enterprises haven’t seemed as enthusiastic about generative AI and large language models (LLMs) lately as they have been in previous years. The Kisaco Research event I attended in September provided some reasons why.
Current gen AI processing far too centralized to be efficient or cost effective
If there’s a single takeaway that I could point to, it’s how overloaded data centers are and how limited edge infrastructure has been when it comes to effectively reducing that data center load for gen AI applications. Ankur Gupta, Senior Vice President and General Manager at Siemens Electronic Data Automation noted during his talk that “the opportunity for low power needs to be met at the edge.”
Gen AI-oriented data centers must handle an inordinate amount of heat per GPU. Gupta asserted that half a liter of water evaporates with every ChatGPT prompt.
The newest, largest GPUs run even hotter. Tobias Mann in The Register in March 2024 wrote that “Nvidia says the [Blackwell] chip can output 1,200W of thermal energy when pumping out the full 20 petaFLOPS of FP4.” Even so, Charlotte Trueman writing in August 2024 in Data Center Dynamics and citing Nvidia CFO Colette Kress, wrote that “Nvidia was expecting to ship ‘several billion dollars in Blackwell revenue during Q4 2024.’”
Edge infrastructure innovation is key, with significant spending planned. IDC recently estimated that global spending on edge computing will reach $228 billion in 2024, a 14% increase from 2023. IDC forecasts spending to rise to $378 billion by 2028, a 13 percent CAGR from 2024 levels.
The research firm “expects all 19 enterprise industries profiled in the spending guide to see five-year double-digit compound annual growth rates (CAGRs) over the forecast period,” with the banking sector spending the most, according to CDO Trends.
To justify this level of investment, Lip-Bu Tan, chairman of Walden International, forecast edge AI revenue potential of $140 billion annually by 2033.
Aren’t smaller language models (SLMs) better for most purposes?
Donald Thompson, Distinguished Engineer @ Microsoft / LinkedIn, compared and contrasted LLMs with SLMs, saying he favors SLMs. It’s not often, he says, that users really need state-of-the-art LLMs. SLMs can allow faster inference, more efficiency and customizability.
Moreover, a solid micro-prompting approach can harness the power of functional, logically divided tasks, in the process improving accuracy. Thompson shared an example user dialogue agentic workflow that enables a form of knowledge graph creation. A dialectic that’s part of this user dialogue flow includes thesis, antithesis and synthesis, eliciting a broader, more informed viewpoint.
Pragmatic enterprise AI starts with better data and organizational change management
Manish Patel, Founding Partner at Nava Ventures, moderated a panel session on
“Emerging Architectures for Applications Using LLMs – The Transition to LLM Agents.” Panelists included Daniel Wu of the Stanford University AI Professional Program, Arun Nandi, Senior Director and Head of Data & Analytics, at Unilever, and Neeraj Kumar, Chief Data Scientist at Pacific Northwest National Laboratory.
The prospect of agentic AI is placing much more focus on the need for governance, risk assessment and improved data quality.
In order for those improvements to be realized, AI adoption must wait for organizational change. Wu pointed out that inside enterprises, “Change management is the single point of failure.” Even successful change efforts take years.
Moreover, expectations about AI are often unrealistic, with executives who don’t have the patience to wait for return on investment.
Kumar underscored the cross-functional nature of AI deployments and ownership considerations that arise as a result.
Nandi figured that “70 percent of the effort (in enterprise AI initiatives)” are change management related, and that such initiatives imply a need for much more extensive collaboration, given AI’s cross-functional nature, with the right people in the right roles in the loop.
Effective edge AI requires a different, linear modeling approach
Stephen Brightfield, CMO of neuromorphic IP provider Brainchip, presented on “Combining Efficient Models with Efficient Architectures.” Brainchip specializes in on-chip, in-cabin processing technology for smart car applications.
Brightfield asserted that “most edge hardware is stalled because it’s designed with a data center mentality.” Some of the observations he made underscored the learnings of an edge-constrained environment:
Rather than stick with a transformer-based neural net of the kind used in LLMs, Brainchip advocates a state space, state evolving, event-based model that promises more efficiency and lower latency.
- Assume fixed power limits.
- Lots of parameters implies a lot of data to move.
- Most data isn’t relevant.
- Sparse data implies more efficiency.
- Don’t recompute what hasn’t changed.
A final thought
Much media coverage is focused on LLM behemoths and the data center-related activities of hyperscalers. But what I found much more compelling were the innovations of smaller providers who were trying to boost the performance and utility of edge AI. After all, inferencing is consuming 80 percent of the energy AI demands, and the potential clearly exists to improve efficiencies through better and more pervasive edge processing.
Maybe we all missed that strategic partnership announcement and thats what drove the price upSeptember Tech Winners: ASX tech sector up 7.42pc, still top YTD - Stockhead
The ASX 200 information technology sector rose ~7.42% in September and remains the top-performing sector this year, up ~48.53% YTD.stockhead.com.au
Either
the author’s source - Billy Leung, investment strategist at Global X Australia - knows more than us and leaked it… (rather unlikely, though)
OR we all somehow missed this recent announcement of a strategic partnership with a leading automotive manufacturer… (okay, I guess we can pretty much rule out that option straight away)
OR Nadine McGrath and/or Billy Leung must have misunderstood our company’s recent social media posts re the advantages of event-based computing for radar/LiDAR applications…
OR the alleged announcement happens to be some sort of hallucination in an AI-generated text…
OR ???
View attachment 70335
View attachment 70337
View attachment 70338
View attachment 70339